var/home/core/zuul-output/0000755000175000017500000000000015154535010014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015154555650015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000336021215154555447020274 0ustar corecore'۲ikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ ?ȋI_翪|mvſFެxۻf+ovpZjC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;F}Zl8T*v (6pk**+ Le*gUWi [ӊg*|FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿/h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߿)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/Ζlpah E ur C&`XR JcwB~R2EL9j7e\(Uё$׿atyХ?*t5z\+`/ErVQUxMҔ&ۈt.3;eg_O ξL1KiYLizpV:C5/=v-}҅"o ']쌕|tϓX8nJ*A*%J[T2pI1Je;s_[,Ҩ38_ь ͰM0ImY/MiVJ5&jNgBt90v߁R:~U jځU~oN9xԞJ|dݤ߯R> kH&Y``:"s ayiBq)u%'4 yܽ yW0 -i̭uJ{KưЖ@+UBj -&JO x@}DS.€>3T0|9ē7$3z^.I< )9qf e%dhy:O40n'c}c1XҸuFiƠIkaIx( +")OtZ l^Z^CQ6tffEmDφǽ{QiOENG{P;sHz"G- >+`قSᔙD'Ad ѭj( ہO r:91v|ɛr|٦/o{C Ӹ!uWȳ)gjw&+uߕt*:͵UMQrN@fYDtEYZb4-UCqK٪L.2teB ˛"ո{Gci`du듎q+;C'16FgVlWaaB)"F,u@30YQg˾_YҊŏ#_f^ TD=VAKNl4Kš4GScѦa0 J ()¾5m'p/\խX\=z,Mw˭x:qu礛WԓL!I? xӤ1(5AKRVF2ɌУլ F "vuhc=JS\kkZAY`R"Hr1]%oR[^oI]${&L8<=#0yaKL: JJl r;t#H+B|ɧJiM cm)>H=l}.^\ݧM<lu Y> XH\z:dHElL(uHR0i#q%]!=t_쾋-, vW~*ſ/,e?IsoSrm_7dPΣ|ͣn/𚃚p9w#z A7yTJ$KOL-aP+;;%+_6'Sr|@2nQ{aK|bjܒ^o(מO80$QxBcXE ء\G=~j{Mܚ: hLT!uP_T{G7C]Ch',ެJG~Jc{xt zܳ'鮱iX%x/QOݸ}S^vv^2M!.xR0I(P 'fΑQ)ۢWP Pe>F=>l |fͨ3|'_iMcĚIdo阊;mT!{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3Dc?lm$K/$s_. WM]̍"W%`lO2-"ew@E=$BrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKd9A7޵6u,? y| \;/"ΩeߥZ}r5d + kg+M0nCj.&O:g\qwV 2;g?;^L)$|;{bH)F[9} VïCYJY2+fQVUSFZ!Gmȱ 9"xLg,i@)2VՏt1tBh0E zכ1ش&cJV$dWӊ皠Ys둴.)Lv.d1Fڪ&KZXYIƇ?{wMר|"pI4%ez1)_zLz"`Z&Z![0rBK kzxG|d78[VZ3JO'| { Qk^K|h>f|l.g_v\TBZy{X-O_vwrpP40ޢE[A͝ Z5 om2p)lbłx_1+=u).s¯c<:wN*|KSlg_pvgIxpI"O c S G-V+Z:H2vd,P4J8 DV|oV1>a$]Nb;s=Kx̊ygbE&>XYs䀚EƂ@>n>GgѨ@OĹCKØi4A嫳B@hn*ӗQ7 }ˍzR*mBA2,'*In?K饸+ wCV?j@)MΈ/\)'Q'8 2|J0êϓ"O+mmBzp-p޴&Z'Y-B6-[4 u>r,8?>7uCC5F %Ն R Cu8?28¢E We; P0:"nlMvOezR= ]â. U`V% CQX v#'Nv%j1^ܒZB$*c.)H ?`5[Z!}nliצ) ݆Y~ cPwίX"{!9V0~\`#U8V$}bpNU`ZS̯窜8L^O+m-q6E¶8^ SxR1Xj~=rs>NdMLmtїUJ8" kcMlf2ǽyWLiXCaSmMÙBgr7[ Nʇ)bAgX'f0]agB-:YSokUdE6c>Ql~JJ#`~#Eh3ŕs,|HrVh7m]Q!ӥSVB l)DzJ |0M>8l WIf|\8U*hг˅r-3'^\  [Cr: QvOS}ll>ŰAVG YңK\rnɪq(u$Bk|TUН}h!8l?= S7!8bdEuK<^F hxndSD,Xt]1Gm*h%)(=XUza^&6"BzξH( ."uw>.,SzbQ!g:0r$ضz['.!-r"1MCMu(kP|еcLqr/Hi%(azry !5yY~ :t>gY #)v緢!BY)Hc(<|qDOIW^9oJb^;sö!`hDZoe@JLw9|fb>ѺRce w0/ qo%HWs\4ݞtf@Զ߼/ȇlCf0C`~ ƾ}O+D7 P=xD`^_[d0Yݎ@z BO2۽k%xe'NԋKI`Yu>7v$Rf[A'2)b>~w!I.WwϏߍ|~x(qUߤ^ӵyx^TE.7Cu/Z6m.Ǖօ^ߋӽfz./ZܗWrn*up+t\OW U-C_wS!|q?E-S_w$-%9?nwh{R 5mZM{=x$ FM8G0Z'Y-X )@V7G)zL~dپ&+<вFzr?㘟,q~,b6)hRgbiRgH1_'9(ɟSYpŘ-ŶםG+qY1౬CxAX4xK"5XoITd E$Z[NdwS:֢̆ ?GЅ'JƖ'ZX t{O֠U sVAqFJc,Ih: ݢkNwE 4eaK0Z / > <|d= }L3Qf xdRhEB{s \ d5;*Ijdɥ(ȊV£55N ۃm-߂8 `6w+ x-f  E\ڈ t-nɭ\h%^wWt!{q.lrDsΒV I-*F~L!Gf"Lf0OWv?"1] `W+7m4e逼p +{6g߷Pg%,IR-Ř`QbmүcH&CLlv`ScivG'հu7.c 61spH5SCt)eNqǪP@o`c/#rv6*;WJ;[.)4R\=V~ׅ@~>_)iqT 8;DQs@4¤>mlN"jek-R9~ {]'##AAwLѲVdJ.ԫiE׬܏ȱlR~voWP~ RTM#rQίE|s`poSfbpnVGIGTq3JC`,$P_KxB[cz)ut0F>v1AWHй qPq3TN9W⌎I Nʇ)bHb.uD`X}nl}\hSؗNTlѩyۢUdE1LӰWlU&Z!hoҒ"HgKX 6 -{zwҌleb}nltwfCEI"*o'Cpl0\neή]kY8Cgcu1!_Ɩ} 8$ Ўa 2+ም+A[QA%Oع „&䒿e;]0:|IIc(R&#ӓecE,dp&nPS '4쾔h|OO-D#ڥ4bZsH{ g3JDcSKu9k-nUAG`2y*8F*-Rz_zҗЩsd6|`U>F*KZmR;~:xI/iFIu]dsEGزM+ycF;bH,te[(6Fƈ^ʮd3 b`[c1髰?(o$[e(l^ȣg3K?nEq!*RZMCpnse~ Fzq.8vz;j}ln2e UZAZv+QD4ͺ|@lX)ĕ̽=Tse+ϙK$s`hnOcE(T#|\*& >Vv:V^4R7[{u%[^3GGL']9~yL :!Szbvc2.R3`Hz)OfJlӪVWF fK6%OOE-kmW`l[ۯc_^ ޱpcoK{[#U jt*h2V߳`JBNcFSUȿ1:WѥLf2HHh)b>~w!ޤ&\UxsnnqCq]w߃뽽){Q=\rw*}-Om )l?7zju_݋yZTuusn Qww^"W|{_ӎ'|;.n@7 >I m-8NJ\NA څB}>Mh/A8?_;qŽ6xmPf_Sp6aMUAP*tLn$l{=lq&Ϻ/j璳PܸRKm;RCsT$Un2@ "G{}ZI>U֞ = ƹ%>Hŋrw/t:7EdE6ɗ%RMn7s49۝; %RndQ'_@Iv4)Cmܨ%@p `*,bO 8pLOLar:xy5L!p'7eyr3L%T^U@tJJ񔔭))[a 1eN,+|i|/QGŕcy1renO,c ;*2cbM<ĵω(wC0Ԍbn,iC0YR5+F.jnfvz#9r~+Yz~(ale<xa(jǓI,2 kcYQ?_n Y1w/I?oͿ2UxzIzr#!>-sS 1mqpL?00"2`NG>G'9' 0т7x܈Xw(O]98$ΥI.}ifpI15-{횿Dh&T+ `j=`D=כGֿM\B=ף/!o<<^KXoǏtRx_"J$T込1KcUbK"aDl,?f+ J \[]x)2}wP[  dqw8kybꪙŒ$&nL"F}UIaVBNl`XZOG%BDeuΪ~R/ƥކSXJl{ yg/m*ĊmsY>umE }jvĂ&v\nٖ‹đtO Y:u~FA@uƞ'#Eqe4NF,Fsx$S2~ty=_RU=,?J? 3cJW#կ_0>`!rb%dj(q'h_Y4 @:(l~ŀ`kzT<,o(lY)ƣ秬՟:ͣZ?<),㡼6 ِI׾76M >WSX֗ydqrwHyȠK#xKK9+O;kU?Zs=w.;b6^N+)5Lyi#ףc2獵"PsɗvF%V@tw'Y&x ]Hg06uKmnuz~6rLJ{i&˪J8<ʤjN|o쏓651q][GAv|Ip>Z6:!.Mu~ ewKzLc1}l]sjۃCb$**eКrdDt#t4CMksbXp"ARҺx" M1˩9Ш<$N"&X4Q[DAAs,!/;/ud'A ]4?S#|7Q#7%F?ʣT}FQ:b*=@ jP@YP=%*Q(Żs\&G/ۍPD6ۨ~Ѳ|smQP\G_ q 9n=@Q[? 50NׂuVXQ讻f2˺4UQb]˜b#RW4'e"em[;q-™W1KKlowF3E?< s!YϏn WE8#xֶH\= uY@@C|_{m2g<+_0CRB U Y)Mizh°((.չ>a;,ZG`5lh]e7VO<wW}d [Ș%әQ`d A {[N]\PK< e#۶|h<žZŊ='}HlI'6ӵvaxW}%:ҥW,o~ڲ)Rè.A&4u!F]~U4 &jm˳.hx2YTFRʴ[pf+8вf%p܇ DӺ."zKN8tK%GR{+qSϳkoIw=SKɺa $EM7o+Rw_l'r,`]IF]$l">tij$eeEbٞ+[UҐٺH@V%'Tw@ٺ>]Z-&IƁK3މpslkCAu%沸f|TWl}xbS\.]D6IxxxQ {O"f4 ׾8B=9+kN*$fQ߲ǁ5tʳ(޾&":s `r25Ajä,.f~]ۧQt 3O0{2,Zy9 b_m]/i- k d &IРb$*?~8QGL|1TXF*P_`ofQ̘aXgl.>5}qm>/:% K?Z9cjBtbz1 Ԟ}wS,(ݣ铔7)!O3l tN8ޞLĶI S$lc- =nuf-'ٻ޶q-Ww7E-l"M̤MҙE Kr-y%; iئ[;ԶD!ݑW'soYt}}ΫQɎ[6^5^|E[u||3+\cx#8'4E8\dYOo䜹r'N ?ҳ_߼WC<|,8火JȆ"YQ]NN1$8K5y}!⺎K|V*/RXeKL=߳5%`ZNj 4/fTH{[h:PWZV<0CqBrt5ȗи=t[VVe+G1lUزjӣ>;p|!~]yutFο?ck;0T|Ag؍-: NGkwtiN8nFJ=3pXGD kvE0hxu4xaFk@ Xz =@8G x0g ;:湖5v ow ixԹ[، @yɹD;^:tαzEwiۑP?R`w]`~b=k.{ <݆#Cޓ'6h~Z{V@m,wŀsa\Dt* R}AHC(I`<ڑ8bӀ]4F@0w0q:hHN$Ny;$SB#)82u`g]u-d~w4aMi;|tBj~GP7l2qQAL+P, NÀvŻŜ?4 ;jXvv?L[yCvO 84l?3A +Kϗ+s`q;d q`o`:9#N 38˧Cv]+`%{,y_o lPFVhwkxhFNy]&urY`v ݧ/;]8H@:Ƨiv PԵ2de tůߧݥoO֧~gҮsnjъ>w 8AG]Z 8Y(_:QԬN䡡Lkr)Rb yM@aAJqr]Aw}v֏pk|,n?G0TUY$CcWSR 57%dWQ>|39ROr0yLˍ\w #/ 2*3"r0>&*Ӱ&ӒT7II> @(5)su}ZȔz!z0aL"êLwS ӫ$= U*0{+L+XVʛfԾ }"YKf<ܦGd%gT/Sޞ3Yc=Jw|*v$Fg«D(82̎|ILW3|b T ~6MsZ<j+)Z^ l*a7O0J3ze|#y <w͊] UuNt -"I\O,>`u戜q|2nĪ,MEӑ86ި*#[pY/?L eaz6FȒigtRQ` }OyT7`N(н'䖒Jh+ RQu~/6 S9/tWK %f&$z&{:tzZղⒽ`ȇШ3N5:W#k}º=yǓ n ,KL2s`Ǟ8H1׏ivۿMcG6IImrL+4-K+| 7gI3e:&\XY@5b J&>m۟uHI@: ~\-o{.DE(gYE|2#t<dG~H%`&d5 8HTd7dSgu5b5 j~kRkU @|Fȃ(}?eLUj0i-QXȗ0i Z< a~@ ÿ[=tJHD-b4,]=iˡbD}$m;0Qͬ'4WD/e=k9O >MW> gFvWOO37􅧧7tЋ J|8a*o ?xPGn!BسcU֦c/J(,VL.I GඓŝVb\Y2܀ml~?m3ߓ*ɯm6(4 Z8*2,"h+P24`p_*}5EB J7$CҐuƔn0tQ+]k\vdhۍweԲ5eʶ#mH)[P{P{ B #ސP{ BEB5u'َPgCB5u u ]Pw;B u [$[Po}B6$[PP B #ߐP BEB5 '4؎`CB5  4\p;B W:DKx DKQZB1E=5tyE.(>Vy]Vh#˧7lJ}'!a|ߐd ˬI\eHJO;SUU=$*e$V_hz(O~ӫ^xiDfDyYpapmYs[nK # \jmsXu] S<1_gx61Ma7X9-\u|z`)ed8IW0gLʔجh9| ̠-fYch/d㬊G&W-۵Ns7折4xΆ[f}S? oSv\p׬/ofb( n"ƛKX*N2@u`y|,z @!|n ڷUr7;턭!|n|`W8$=@@w>|#uQN2L",; 5^WZ0{un {ق Mrb>*ӳ<9lzrva o/xO]KgR0@mSٸ )n((O&0j6]۲UL) hmiPT},l‚C%xzsoS4լ.zX}~:GgY~v(d!;tMcj(i_dʝ2~D"G^2yamP%ܥV}GqR㶎P\LWgZ*aM 8)9IP$\pjP$(l~@Wh`ʧ ڎX9ѧOosS{ #+LkM.9 !|NVD_8x [C;")Y˭=H8 ҋL9ҕóO#ܝ]!8r<45M\G;߽o\{~|o2jC~FG/1'gceα+/^lv_ ç}CWl:*g(襠ѓ6rZ&LW(O;'jl#~B1S8FQ8 U.tK:2ىWP$M9,Byṕ{ MŮݚs zVL/*R7[=U>PxϣZzrdM(JEԡo!R@Ϝ_w,8:5G}pV׶3i8R$ő0A3y`4rpvmrzEC"Dv[8ptN#Y`jLS (z"8`T쒖n! dAK'KqH<IGv  jF xSAxۛ/4MҼ1BS>$(ǑM;OZp-?Yr"$߃yI'_鹰aݍrZʗcQܡc|eM4ߜ1TΌ.+ !VjA XL&O*/ 졵BxEE&$q_N#1|0w AƘ+{i)`s[ClbFV Vs~I5lxe!{0(ap~q hO pAZd(=H^9GR)i*8*|se3 J̌\e,9CW cZt!zk0^$K/4UÚy\NhVۏ[.CZHt㧏%好&uӴ۲z`gtu,gϡxm6v04RN]sKy !ZNIw}1~QU49bGMB:kAhb<['ȨCѴ6Y_16LLe֣s;8:O2|`b܌bR4vr g?2 mFfis=bGJrv{;ٌ;LeW0;%&t8*:lZԖ:K]yI'!n+:1Gmt)`ЀQkm)Npև΂-((aLɢD):?ః)ϏqJx12q6[Ԗh4(No*?89q-9Ҙa;5|6o{AIiSF'/UIݿ'# !0bd@S<*xᬶAP2 G~a1J?p,4%b;ɔ Z]ǫHh0X+/jsp{$n\_'s̤ɣ|:.xп]ٮ^E~{f#:Ə\#N B7xDn]TE'ՙ/VdEx ~4 F@'JVh6(NK>y탓?\v `w!}di^#sB)gܮ`uWkHal`1/jnhKHCv8/jViP9Qc1x,oQ)!uMˢ2 d[/k9c̥Lm層\ ݧz5]S܃ċbyDM N!x[hX"[MJrL=\lr6BTfӳ`j\/J;E|^?v9dsb2\;wTJe0h⨨/-pIVrt 2!UPZ *p#,SkcM([,^";{Y;]-a2}^9J"%Q{jEqVc]oZX"4nCpr$N &7z\ϩ\'}wo3~< WJY8;|[ R`x1_h@tM)SwDE"tm9_9k,sDTvӣT hT;8+UG5 @4I':苹,8KϨJԓЭ9b "!u|*K xsT aPɃb|mc,upu~_X쇎2 uJdv[9p4]G RF*7 }(WX4롓#?xn6=`]3FsݝEGqwNzwςd?cs4ި`8ZU'0)v6*uJ+C; eGIhtnr\q@*`# 6[D7ɆϞح>ܯ{>nvT_QztujJcieՒD "@.yc~F$qw>ܧ*>di>m?'zAfdA{hm6CT&x~krrCm=}^ /y 19Y͖)(BmA8TRwJFIH:YX6\tR5bwE҉s  2: VTTnIV%){ys.9PrxEGBڽẽh j]F4F2 yc.XH:ߙ~ǂdsG: 8>^k˂EU.|E?W!~*\~Dԃòݬ[}J39#frn IܞE#Ϝɻ]K!gOXxe|J+ueAվ>K:q!τu0F\T*y%K7wcިNnԐQqN'j'}_k[Rׯkl3CL!YhT$2؆Μ0p]?y rͷ棗_X1A_E8LYE๹I\ Kx;Q m*.4Io9s7"HGǁc4z֔{4朽M@2%:EIu&e NXS' HjԻbZ{@w"/3Z-`˄ +f?ٰl7}DlpN&'995Obc$vQZM9їS'ChѨjg hI&XQ$8۟,8Nn;xIļ7|\>Ȝ;Գ[ڄ/ cb$d1֮g*1A"wh^TE5<#DeR4b{n暯c3dMk=y %K{z2?% i䄸 9㫋jC_PO ǫ,8N [F7| FFKH%4ByEKđ"ǝ`FgTLm2-cs[_Mj4s z.G48K[~&S}Ra)k;9>NZzc݈#z-,Ź"" pGLwpD+|kAnXp\K XUn cvZ<Ą:Dע1^oYp#\ ҃^rW|ד,ZǸgAHI$M]Vy=ẍM{U?R\9.Dj11HqQk~~5?o-`EkqψwY.5x5oЕRypr,+DIa9p'O8lVabY Ls4(u/_"7 Y|8ܩE ejBٛΜ]0;SfAݓ<̡2ywmYX5dFGkɸ&)U ^Cv $Hrv?wE1. Dj ǕVIҭNoϖpưC1UvȜ6 81B02h"JK7 p_82`-.Z:4XC ί{fDfOr?v0w`Z IY~]YϏcD@o/u#Ɠؖ_lŰwMw}30eߓ/F ` ܿ'ygg%y! O&<7AXC%1 ʱ !)#8Wa|I ;ϓ)9鼁+2ixG/Lj3"eq:oLǾ/QːI|8EQd^l&)<U:Nv;>Pum3F;-!<!"l/NQ2{y GP7k+ 3f~EQެ{6g,Y %0V> ?7Nn\ D#xr ˺L\lrTS )H7TQAsYo 7ՍflgcdQxmb<dftx|-l\^$惟O^}7e7<>aOii5^ŗz. 冇 2/q68e( M"u,$ޖ/g|w.&Q=>-xG@xv]7㨌? DMN~6d8[ N7q0N%Ta8h^ KOup5 1W 1?.L\NWp\ ڏfb/nv%o  ,)mY4YwM Ҥ;NkgRpäifR.inG xFma0PYꚃM+=Fv:Gpx/_製1UV\CR~PV`|k7Q/p{g&Da3*9 @lzg flKshm#W UsXS(rb(;7T$Uu *UAߺ.-Ԋ4 &a1*s Nb܈ gJޜJ J-/W1}䬳M#TA7 y6rpphѴs;וUMq}8UWi反[}SFl?,'߽|IJ|҉Yp\darb,1>X|z$9?>8Zwݸ񽫬]57t+e_92 w~o<[}irj!m*(~J9`!M xveQf`-[7rmɛORY-)ITI!M:-  )G_2o <:ZfdtWH7pU۷ouc T+٢nX?kFxTu B0IZ-S=BD*[% PEUA3*ihnA""iHIRܢ'bɛPZbቻH,HK(NZA(}ZX;/u4jEg-AUhQTWZImJTiiai H.b Jq#µc>J0+E`@0mќ`;"V7F+h Iz>f$?H>f^2^~/VG90XNrD _ocp:9(oTNYԂ}0CߠeIљEIĴWz;%q._\]1ڽ?: %R ;-n| ϴXօ; .0#Ҏz4VH-Sm08eV+%o sN^/k%~y[Z0i>> `@/h/MדjB:j#TX0fRo'ȝs9p 9## ;* +ܕqdYrdwY` YWpl`@dBjl;7Z~\L;dR\a[=kr|߭GjW6P̽@q8\PBH em 9s(V{?XJp<94wH)ʭK#h(Λ%m͛T}FNh;N&fCͶ-MĽ~7k[_͑kcL0-qN>s`i}!l\ ۰Zlx,Evg߰f,_V1LڴE`yk"5ÑZ|=.7گ)]t .7)"^@ߝZ_NYq&4Mp]@Ko*bSs&@"%<'Jđ8 _vv%#Mj̻groae7Ӟfֶ9 OT/H.h*¶sǺ`Ŋ'> ojߓt3xVO@mK== ]?g3kr`iBk"j[sŭc|F{z[jVm|Ͻ%I'#ۛF}}7F_WMN8B_=n@bxa.bU8VՀ[>!H3zp+a \XUh\(Wӭ MH2L:`K+;K/we!yi&4ͮt>7T2/Ϊ,$ |MӖMϴC,\ W#|sP=2|*h6p4B 7XuP>xs.vKw=x;?"L.b-?j³Ӌa4ꩃ >C7|m>1Q "=Ob"B#fBK|@#/:.>(&߀25rZ~c8y8v]!'ўЪ֏PzJdSp If#f60h[f˽ Rz6zS2AgWG*+2!` jO(/[[З}%~mcЬ|u*_-RIhmaTrA@!gZr61IzIJ`tQ *W7b#nPɩZ|&W,l!c4 YX0}~dq/1:sS*偪TLJ3ЩSxi"*)୚6{3Q;y5z#X|ȥ:RIcs/X&9R 9I \:G YģTpHCE%' Pf`C(~7>$8yrA@,|ʨ !BcbK2H* ֨U A޸!u!Uؤٻ6+Wv1 V^ 2Nb*0kGRbjjYnel~^;]fCKxN`9X( VXW Yřu'A0F X`Hxy_d49Ur+5f%tJE2 6-.֥Sܨo$W0/UlS#8}'ʶ$_ TʀoFl`5pl VpWfEVf>N5%:c{B}g dR^Y!B 8LW3c+1F~R1 ~ᖿ x[Slm)m]ljRvȮ]R:6IZYdf@2pfU(,3vQ WR1R:Vw@PֆZ`.#nLxBߴBZaiЦpΦ2hRh2( ¾Yj2eDL&xl }6!M;IqZEjWVjrJ. UvVN["S&IJ_^`,3.iCOTn &YD-j`:‡KNkQc Ә֑R3Kp-ENU\PsbtRJhQr.$+nQ˷Dv{k}يN+8w?-Lνrd~2-(հ l 6=m$Tv@Jh'`R'k|QҰ$D4/Z1h 9 y` ީ` uH6p4n'Ob+lxߟl sz27N[UbH&aW]vqn'䛯L슦˪ᜑ]{+Cje3d( 7raV70ɨՎWr5Lչ\\[*UʀWai_>RQaJdcqm~m~f#Te^\s"(~9 A\K])n\-I%:4tǘRҼY崀e5hw ^+Cu&Y񪤅IntM .d`n2&p_Tܘ֩}\w  .Twvvj7Xv/ pNo3^Ľ~Nxi PrcZ^nS@*4yۊ3sܺ0;'gв~4[RLlHG\W\TTdM10-!0-H:\G\uk7hݩpwJ_dkyNQ+h#.X @RזZof(Ŕb>WЮ6"wdZH ł Z&P0MWS>w )XĆ RSADCSD- wHVwP0%$E#8lZŘdv_bHf֟,ŸQC' Գpibu"Ѳ#zQVdh^Z(Cޚ{>8G&B{>"93~(7o}S~q7e6 r:Ѳ}Q_֙_OZ*"#p,ikCP̣cDCYgyqx`>x "`rSO\<#E-l%{#JGpk9//fPqX¹ݫr>;/xX& |\ޝlÅi@xvzŌO*9軘7X;(V{3o^@[@O@|;x짯 aCݓ3Y߲C9|DtĥKuƪg/׳26BZk4ǯZ>{<<*R%#BH_7W~u.u`R?qZ)&yf/9ڳ}̍UIslnM3XShe_u|"}o n_>9K߽{={nj)o]_r8xwfxWa@K4p Ɠ,h39}u9ʮWw[pEdGx YMWp(xK/5e*iw|˹?Ns]kw[w 6k]JŞr̙doO'S \o/7Ev  ٓGS?=n6Mg~Rsi1)+8Yimr#g]fw ¹k^܋q q6a S:'~~؂fDUJ|0- 88 wyu~ߗc/?^}_|re7vUM1(8l ¿zL,  ֋Qj52aPp6+>qB' GJVӃvW[HM.Jw#BS6b^{ǁPq91 jlsh~)Bm6<͛&`uz>޳,q mIuaw}}SGN(iV$,+ْ2_Uy=]7=&G[,Z(fLB3o=PaLrH~%a*ѩz3v=?G;jV &} 13;+v [եݠn`T> ^oS]?ݵѿtu#J}D٥5h䝳S8fp腢t@v:9]iW%qy W)tI #C+ sQTV: ((;loM/ΰDp0mieI jkTvQH[Ԇ2&҈.MTr'7u):-eo&ѶVkOPKڈ''BB` ,j~0z~J|gj&@++~[U9x8%Ӂrdf_:Fsw묕φ{.?bjdWMilԠ]ݍ@,w?r?ת1Cw|s5SFH]v1̤rW00%I?J ؤppmb V5:$L P[JzꫲDh&%G]s:zfiº<`:<\jdciՍ۬,ךD3 P0 1ɻSZԍf1h-i$5mIP jz&1 Uwbx3JB3?$E0 4mt]Bm^uR_dpMߴEtC1"oB4Pϡ*ii,(H~0_Vd0ue\:]Rsoe[RF_k=VT 0< 0#&2Ugl8N "!ԅ8 R3{o̶s&=4f|⑧@8},{84Sl;¢ҚLe``*ahrXJ65#6.1F'}4Z|}KnfX5CFK#kɝ1taB~O;.ad;F^{]ӕow8#T\ݜ>A~UT}~z?K?Ӗ!=kMͪ'o,nIߖJ8BEZ;~cӑС08~H~J}"?Q\Dq~ LCu8? 0{8RDin E}9vD2<>qLZH#) b٫x${(FQM`HX~hd= E~X~t\*h'ަb"Y6{%g!EX_bTF$S_If/XŏPZJٗtK}Iw[+'#Si;};4p)42#pQh[X%wzBsn@npeުD;X3ON|!n1Hd56o ^QsƊP0}kbEH&qM( .'D$ЪGcYf~jO xG ;z|?nVVTi+tZ[dPYFJ8GKpWB&b.UPkʊ D0v𼺞:Ӧ"Xrw_$ʭXgqgqw_ #C`-2mЃ azԋ~Ƭi} m}Yqm-o+‡u'yky8K ؿ}C̄q4c;03qVT|;ӸsYPzjɌ +eʸJÁerWFOJ1-2Aw1,xtYp_tBdUG>S砯Uתj-U1TA:ϵߘr~ʞ Lt*@6' pplU֮3y9@0G$ 5'̚>|LZnYqIC%{VZhQj˄eFN(_]ou> ^g~o 3V«oΖSLr>'% bW'W8&V8ƦUҨvvԗS%d/)&8!1oKk{L f`-rᖷϫ[.5q\ҼrV2R\,c"#lE˪JF%L*%IEr^Ue\kFOT SQ ~T R5cM&QiM ZMӏGFCASvl"@}'FU7E δ:}bG ?1MLoUcUt<["3<-φgOS3:$ AD`GWY4L}fujo v`VuL|,;IҰ@^PGuLa`*MB6$k,e6;j[K58vb-pSp{ry5ȿ9"`1zD #0(}"3B+PΟېTlhoIeof#:(%f<r}ݧ8?M;|명Q+ͱ/jb`ZJlKa`*pmβ`J4kEB<=Evb.]{<5ZDNNɍ,eNơU/Ev8 PK&wؐlr#3$Siu}D2\{ ) zf_L'~'6!Dv8|(6 9\TZؓP"T#[~jHã{zwg^$7{)HfH2#H1D>3{"$P$7Az2rtF7M yr RSG(s^]pS:J=aCg#F"i1 j|J^_!R.-}?v+]=lYYnW%% V8lHA(C1"=o(QzUɮ)eM)_V*RuæyJ)K6d."UwR1*S$e;! 2jEdJB5=4͔H!1%Գޣ>8!UT"UWTaj{zqRQ؀R]ouz}LZ\Hcz=VKH&_/UP^>wؿ9+Y~^y>1?s .7^>靡6EOsCVݣk ƶFZ\`8aM:mV/$cFMdZ <2 OA{ލ&\;׬k8ӿN=6NJ3IW_w?}rh!:k'dqh6H9ү }|w8UDr4#"=?;!<饓[uqDQj:rCC;mlJ\%p^{. ƈ7a1 lb@E՗*e՟/wMskxq{wϪ嫛-Z6ckǁ9S]  t4xG,VK$K{^3WibIG ھ׃tL8<8lW4.uu&\BJTC^fSz&:ғ~}xAb)@{+QQnCdjD;l[$̈.m4$s}h{i1!zLgņ*ð@xou2Z~M4߮.5=gN9.ɂl-4E,GkgתU}i1~2YH Qd G oBwjgUZYRFZzrW,TJȍq,)|Y8Ґ_ma_KCߟUkz2 6YB R3X1/Wcl՘%1xh֩߿R5Ɓ rpd>H9 l G_9z|p/nocX~j*!t ;-tppZ ;0CUʃ,xj*'y'$@$5ˁlb N<ՒtID%+)[]V< URR(! RjtUq\]0ZcqO<b>H1n-RKZ ziOH0Xt<8nHUu4xd3#X `He yʟ Rsץ҇*#ɢ FBAwt>H9QɌ'3 8 u2cɌy˰3, +6G !Fd)%h}fX Ҫ ,fɼ9,K08BgԢYHpTY2KZ0Juk]֫[V]OO1O0i<24ֶ(2C5C ]3i?* R@<;,)r(}{6^eGRie3Zʩ'ZuYA Ovh]B]eVzzߧO}m2 <惰7޹)2^:~P_Cuno@pfqg9Kŗ 媻ڨiݺ\dzڂr:=z7Μ g_A!g9w{T<퉩7 u!M"=(rgf(7ԭRWRX:us` M5d/Lh g%#H9|I+_f]_xطvZ8ũ*ulTdP,<)vjdg\Rm˙m/OOXB%TOL^9NFbdlS:U@}kU۫:3X_d8Z&u`l&fjo&wmW'ʤV+S0!{q1v8ctxj,K_t8_G20Z<)}lhpBR>x+&.B@cf.H1k͛V4ZV49lEêtMSk[L>dVrY~Bh W[ooAWC 4~ԺWwCVFN:YB6<ޔ+0k2>dj6H1ʯ62Lscu(Ə &Y Ϫ)X^g-ߘoBS `Y2e>CS[WVW_PU}k@6V)7|ݦ}}n&%M&?g 倀MY{`N_j.08e-wCDy3 5tsK}?t0#×P]D1"jLugz}M3Py}/}/tB[w $ZSTR 4ؙ6Pvo Yz6.c]Ѹ%i[P~V@%f\qPWwМHxxIE2 Ap5 Fx-4ި9HK,cɂlr @6P&XSHYXv XA݈xlb<WoC,1C!+Ї`7~~2"2X !!!}M«jv2x$S! 堊uT Ѧ Uo -[Oਲ=;8_ R (euPI˰L(΄ ̅G>em(惔@:MP40TPY?yǿ\ۛ+j3Go|Ɣ ؒ/+H1BTlO7j}7qon?`un90kǎOHCK;ׇ6R>Yn=;@3Ɓaqc:G6ƃuھNof_DzjLZva$qIwdab6F9P+]s~3L?|Lx ӢN)=f, M(ԌeGƒG'hC3#rv([`u}c"yf$Bq9#yFbDfOw$ou5n bmdsK|FVr eJ  6)L5Kei&- 91q@:aZ US2Kxq8`O39:RO&*gS  #1 آ2$v.Pdm6P lo<V3#au8l{tfUsKH}ZG#Y㬵RE z:qg~W &~ݽ.1!ލ,),֙KlF`\KH: } jta5`Fr$MJ:Ռ}3,)i-o]#iW)[JҭA*駥ym̴o8N{; Vz`h/h;{V4rLQ,)+O*tz*Wvp%{=2P8}S(' a> R8]AӾuU;v)A[F@3#heR$OO@xщ36 *AʁJpt3A绺w*ٞRFêmQc#doOw'}5A9(G٭Z|^~kmy-czg6뀈- 1Cz>[3}~ &đ88=h<'׿,OKoo⎍~imFtaܽ;~`/+y}\?>uH\로}/JHϚ_3[{}~ޜޯbl%VKV+?dF`5w3a.%M]H; :4?ea7Bx$6;78EBg uqcqNp|?_ޟ>Ǜwƽҡ/;*oX> b蚇7ǻ~c [I?\_}ꕵnON9s'=8R4:8q> sl2#M7OԾli,hTp6>*Qql ta#}50VX c~:q^(=iFvʍ}@W.+-CF7|d|0KdHBCr~?df!FO?,a~_P%ΈiʄY gXkDpc LO{|r>c|^32U˛fBpMf>B1Js~)(qޘs*z <5Y}Cy<]?FT㥤|SD-PY.}3hnoRC:x2UwcnHI "X 9ZBR6|ʷRJ\:PHY)S3bHyB=i"YPJ)ܶ.){3mۙ|K> ,GXB>ȇ5e1**iPf3\3w<}-5s|+ekʻ9sυ̍I,C{\bw%v+m> W֑F׾|^of ϑX?꽻3 #~ďK~t 0s?)n{0>>]ƃ!].@[ؖf}݇/q_w=ݎ<85$fJ\zkCӞbR6>oJ;(aV/=r]]P?kڀwY*`1ˮ;eڡ 6wr;iK7\m09DoTSa@+=nepSnIJ=n0`ÏQi'׻΅[pG. >ٴUw>@T>b!9&!DC.0* UN*%sC$-PYA-+`> J켴2YEgweQng-E`u9IVnajF7~tobRCQAKg|+J[w(76GY1ro{UeJh9 cE8fIzG`'W;?ugn ָ%z:,_W$Jf6}afUlQi@H\?dy$2Ҝk2\XN(5yoz^=2s UZ m9N~Tb_y=rYOВϮy*m9W"Y |ڡfVIvDoJ[HLrqV0 `;&- }d%ZUiK^i48Cy *%6 ,4ew^pЖ!C0'M QYOvZNbMKٕ0ˣK?(d}Z޷iϫhE"+Fۗ]=Z&m4OJ]R5ONWXXzd¨C{QGmSRGm#6h)F{$!#2ra0 `\jpHt|7RLLtWjnhrw熟Km;JYqb,4]$WIj{?}+s2j{\ r|fJ4Q(ˆyM,W % 1Q >K)H,/aBi[7]w_>cp⫽eWٽYy/5۰(= TWGvjqz9I0AxyºdU;5G Zr䠸%# +]ytm.1E#y:pl3oX'c@!uL牰zl!|Ά9 >67ذJԅo֥5,㶬e򛗘hXk!ydl*3"S^GY5]yrKzHL2+qBV0ءy;B#B+p5\<(9_NFR=!'q2:}0 zںq |khhKkcs[4QQx~pzgC"LS.mg/Ư2w ER ^ ]5zol?P &SխCf1L;J  Ԫ6&uov"Nڨ"k>di]DÊ^QUsBb4A* *[UKq]e;\1(oHBɜ~G+ rG=ZdPu,j4ZYz$VJW`Mx WZQl $p 鉼NO r΃De BUKHa͵K+9n)1-d}-hUG4F>,?ߊ4O}G@xKI OF5?@]!pil}X(hBz~eͳO cW}GIe+(.rX9C`IJN!&:eME⣸$?эcR]6:&L2q6E N \QdRQQwSH ~ȩ-ҽK$|7nWM@3QE/p{rbiRhR[hRX'9 :o8YBRG󩴌$|jPh W [h R9'.WdRǍIP?;j6 JHIza?M5+1 },Ak(D *E mva k(aF}1Sjc %^ΰTHN@]*B邩h!Zí墭"ݤj.$J'I(%Zdd*yHJF-THH6 Pwb]!( cF7&6#L+ ³BNP' Ah42 +X{W1lh2  mEmW 4G$qh EU`o-2 ܅N&Vh11e0P-j";@"u'~[byʼhij-XN*Sz9׆OBmΛ!lUj$ λ Aa[p| , Y9h!JM }`)@E(QE20͡³o40MSگ*xRQ#mE3 C.DP*B Ek>& F$'2"E0pPG/i2 DBdcɌ32g,y.% P 7]&"H2 ! 4@$t`y^  D- "IPDRBHψ1>ykQ0TN I!F.A BxƋ(6;8dRP DgRv `%A׃%x&nU_{D"4,$!(@0s8e`sAW AZ a~BZGo!@Bev j#uIΡh^.%^B!Ay7RayE4ͫfQ |E!Ҝ5P1DPPwmf>-403Yͦ`2EleIt-YU\ii:d}xH Z,˨4$QXiRҡjE}Rn'23T0759(&Θ 91pX1Z(MPD !3"LP r0A6<[d;x=<_ Vkb&66Z$<| x A"ԁ#.mQ`қMHg%f YKm4.p1%Hv9Am+/(XBJ ‚JłE&rȼP>(uAy0vjq /"* _A-(&1 v\x$^["3HpGk`QթF,TqE6i@}4  yTx &suYpFv<,};Xj@w0~t!B]Z/NZM,"߁]e #0zh"!UPK |)S[Z%P P l0JIYB (kQr9;zF$ ظN8 m`y$mڈF9HO(C!Y %,@̓t^A$NʀGB8v**iI 3rr0nmB ⩣"`X s qCD+UZ @K! 7ifR㉎s8(A/%(ޚ#rF<>x ъ5rwg сr@A9u@zhK ꑿVk6bXh,@T #f%"P$IKQ' BÿuI^ è5E|CAi,j  Q1J1$bj`-09/|6yA[ITFX̣ EXii|/O pJnlQ%p i5Qz2J2.C)|{_NBlY[Zۏ[fn\3[ؗQj*ϰvo~V ze»^J߿+E- WYޚp Җp*Pp+uH•[שք{vQ wYަp њpe=r"Z-{(צYRj{{>z}$: WWU W+[=1|8(\E?g0_lpJUDRU<dmrkb)(_@44/K?]mwEPv`tqQs_j`(~2p;:駫NG`ꇿ_K)CƸfkBwo lfJd}a2)nsUOZD`p;q_UO.1 >9i?kǾgojsUw9 w}f~qWł9Uߔ7'bS|+Qxe]nB&+iB2g)71'7W3Xqkdf/x-㞬PɞncˇLW@3rE9D洦VR>&SsQթ9qCjۦAN^m*A̹cy{X՛T?x8+į]UjsO??Xҫ«9oO]SCҖI=/)!2+ PK&h~]R@< :XԟhWX /&`*7gGWR|V\XXR8l? 4ieY'p8ove yIZ?tb~?C9<5'uWS@+NbBO+taU|sxd]vX?? 7>_SK&}$1*}ӄX\L@XNW^X*E!CvCzLS>XZAvy$灼VGaR6o?mXɑU::y}V@d^qnv$? ֍;&yHO]6'=欵_dqdn8R8_W)(\oq?m,{O^oǨ~ց}sZmĚ$=So46J)O ʖn+qg"NbD*_>SygN MN/qC5:9?W>7FܵaO?ʞ# X.._ ^FWkyg)S1<XHQ,nU}j7vu症pRpH\˲QIS Fឝ]j?0S?_>K%(iZ/)69,q> Р풺m_e3LONAءªe3UCWSK_~3܁T? Sף46E?`^Jδd] / )q*KB`5 D\\[ fyƔ22'/̧+^5;x2S;Ww[[yN~_)f-%I~.SP&& 2o+Cv[2XlnskkכOlAo 3~Y<39fr9:y^رS)VC'^ĴŤ^MJw:(6M{n0 ʶMĸ^6/*.(ܳ)6|͍ IL$ŐT}& i~M6ňV'T»C ޓ-q$+|"E(k{ f3Bk̈=FEiD@vV#*yvxc~g+vXf'~Zf7$Oh ޶z$]c+1EW mzg紪ROXח8v}Tꧤ_<-x>:P~3Z50=\TO[[[[ƶK=foe z͖Éqh=ݖ2wtϚὟI}8֞a'Oj=>PESJxOԄS[*z_ӹ_ӹw_jE-%<_.Y(ŸG|N(*W$E-FC0_mSME9~G/jjO8)b?WHټ}ҴvUԷn|u> ~?wĐj=(jƒU"HRD/˗u< Vf77by;H;U5|z:,;;$i_&_N 69*R,s9 )23^hJ&Jbw*h>Yѓ}Wg_s%5Sp^Y+տyd1ƠJX ^P)Cm<9c_f9tw;:kAO.|u*AϦitbs@†-v' wkkJJ%ZG kkS{V6ZeRq'-EzfZxƔnIoE8ޮ#2YGk r ]M Z w+XO(ް>q*B#/mAdKH+\,.1=1_ ݰ1X$ LU6l2LۨA%|HE%ASܰ={c;,T jVR0Wa"KH!heY3#Y.= 8-_oM{iS޴Mp))zYt,Ypy6e_û@;"x uy[߭35n7QQŚ}]]{N^槎nlۋM@^o+,pX}njbv°WŻ߾x~}:Gb{ Uoqʋ0lw,xޭgGJ1UO뿮lJs;[i,];_ܖ˛/'Ofz‘n[VonsVgao7v|wJwڢ >x 2Et8&j/X?}bͣѧ%>#`l\5z+0͟Xe['Nkc} B 2|BɂpG}6 M. 1Gd'Ȼ.h}7s:02$4%qM[2Ɔ=G\&䠉,êSϴKR) )k-%$yI7<#.|qeZNFܵXp|CrRC#,)# Q2"WRfKH-@8S>$|qnHyGRH Ọgc9^M :&AS&<Gs⃬muE؊YE%u6%Z\AIz[uU'WEICF/}hWF0sH- (J֛b)cd`]ȋc;Cz QʶwQ>ݢgX3gl3l]Lr>.xEn'Wjm=cTExr[F91ϔ193O[9GF婦QTgcXc+8>FOf^XX-Nk=eR/BWh uL,$l%H. &ӏN6|,9gRЈKPO!bFNM$' hҖ8g8aG ܑ=2c>z$cr9j|Z½,f7ݛy}N}yywc&+M[Op*Sj£dQV\-AZLe橀[:Eû>4-P {KdV+,Zc\5X/ݡy4KZy NalZ[1{˃$:*E+|\E~z& |$QghNf"4li$N'JP.q0(h D<ԏU/.riH3n$ v?ޅ#0x-Ü1SkdWWpFA?Z#]txs6$0X*3Vώ 8I8qږ3 jUJqD*j cNY m XՁbG3DJXi @6$jIy􊥼_^cKr>QcI~қᘢ\抰$<c@R$:5>;7ȹK3z>(N7c ]A3vN1`\O2P`՚ڶEF}a&cpHKr1/ ώv|6uNU+ B-(;eN+nݤ}7) |ȦnKݠc3=z: Tu`M5猶)ŔyJ۫ |¾]KC>)5g4iŹJ4%Ռ3Q2!G?DKra(^?Ҝמ-:K0y:B+%ꦔZ3DY!3}^IO1wl2n?Y7hK)YҌ^ۧg1b *g4;fޔ: |,itζM''چcqaWp?#JMXeA]鬹a`%2) hrK_(ZG' X2vRQ1=}AT)yyyC~KFx4 qq{u|q ꏓ`@ht0*s-K C׫.Gwؠm& ~>w6qὊ$+!B2eLШ|²^cqC{PcI>aEKQP^xB5GW FMv<}=-~w'>D Ѣ $9EOgŬ"0)oT;|_jfg[~ȹVe%-&k (Ḱ)XizeQ&bzT#KT1ylInfqv3կaypvu@h5;m`gq[b=f\A5V6e~IvR 6upm-9ٜ<_*aXrvn&t2@Em uP(m`ո)#!<5Z0Ƶ|P]<97KUcW$XrS$ ALPmSc1({qqͯNSV2FU$H0~Ӂ*@ǻW2hقr:5<ﯖYˌ<רAohУ5|,uS?QhyزIIO$9*PR!*kK"'pO, |,'~4n1ɱ31!' 1oBf s^8~CgurD2zSf W8 ̳}Kb}>1{0$OAZ$+5_/XOXvs.Kx oNYOuUnBc(q1Iһy%2fU+t˒kv$J1c)>a.]~6A$h82)L6=@9N޵h$d$SU^i>Ἣf5Q5~`p$@#I` NƸ7ߎ@<KhTU?7*: <ln/s­#+7[l~/`QCztQx &.٬x+H~` n|DsT nt{ &PDL4KTp'6P; XKP>XIҊ#cd{Ǎ#$w0p9sA&9#FҒ ߯[(A8)jLG+1Qti[CNmO eyFωA{G܆R1\O"}M!#Rcqs 爓L)S:^xXYܥ61YZi^]$?Bc>0XϘ:rw1?F1N\p%w8]'?.:/e5{l<$~5BcFjĄolV8*CZh;~.p4y/lDPz >#M>?/;CRZh8gr"ƨ#5m>` F$R3qC:o[h9帢-Q͉\GI}gƨ <1W<Msڴx̑5\&1fZFRl;քQ%P3Tlh.8.@|DI!'E,G"6:9s '[ǐ s*{˞.^Ӳ9Xh(KMdG5h+J! 9~+*1 y䱸0Vm1 d|TI1aȥ#&CgF%{eJ*E5W3\8lџi^d__3T*~̈ iy7L&8Ɩ)o_rpƑ8bރ<@rz 󰜹5KcN4JrhL$05!(܆mowۭ[yq:d9K;H"ai{D q$J6r仢Q$͓~.pͨ4NABcj㪄x[*aGWMPG,\ T'9CZul5$B%c@hԺGSh- u_۝fԘ y}]%fL#y{$IkF#zW1HD~ahkE`S fM8|3p=\3dflw(W$U(aQ#Hb"aa9@4g@2\U{miX{ wٱTC0Ƅ\j$cK ~YMZѨu ? G}{u4({$WzAY)(fiGW0ǽ;xO!r 9逥H(܆g.:l9b3aF9E_p>TZV(!.lDcTByZh#8#@Mˁv5,t.x{]zG&߬KmePW .0˿"5O"5|UP|6c>V<ɦͦg"0p9mGfDW^n,0ֻƈ{{3)pwӁԘC3K\4,wq-3F!k`}&R|r_]]ЎrE|ƨ{U4L#K9;e"49<߼AgEJ}R*׾H*WH;Ie" >AlvS/EJ;-eA+3H .aJF]EѬ"Sߟ]|X|ꔤ sqԧCӞQ8(H{2OJ/C$BcXo\_?=x jBc52lڽFu1 RcNj2/sQ$6-4F:Qӌ"w1Y"Ĕ";s%A ;z?-xI鿧f!8PIn?6ݟ٧/ XSW4rZIkKoМEp,<\;ImϘ9|5i~5p T|Pu4b^dEa{gOusć%'1 ]~@7PSGa_ -'/xAQ>4ў%}ʳvI/esI+%$E)a[&v27#SKJ%p*|S<-{0Jdsi͑a2M օ*N!J" ݼ}"TW_A{=?f&X0z4icSD& aL;w%K޹}45w_B$N<)[WR ;n,41 ;yklZ9~41jIL$ofN4#5F㌺(F PSb0kv:͋i=e/^[?!zPNZ(F!A$oۡ3H0;ߢdn8,{$4ܽ=~i&< zGQ8Ǽ 5W.U dFs4oZI "{LBW^ZBwxIV)m *IA֒IjC^:*zh,kX%7Zn 6пh42m)ۆ\t#91@9oA"f0n6YUt"$C:iuh} W BuU)Nڛ)3%^-K0T/_?w{nwTUYh93s5rk2瘤Kn]Z̬{BVB!7[j<\Od €7ځ[|*П{wϾtȤ_Zh 'cR  aJ0udTn8 3SSliol([0FUdMCDb-X#޺]#Ohr/ApA}F.KpYDiIR&&]nCU%E84"/e 61ᶸϛ)Tű=e|~ zB"GԜ)SL~\jw(bߎO?OyprZUfoe@Yd: `& ދrYͩk^ɝ+&ä_}3ey0aN}_U(|LM¤"eF@^":dt=3N9a I¯gB[ 4t|]e`ۡ#RPH??_G h8$'ʟ1ױ׮{ ".yʼJJ9n.W]ً~zn6:f#sD4C4]䷬z N6*7- ~O7DH7[-GPz}[x`=,Y"#q+RIܥ޼[$Pxe>;.jgCW(PᎱiTQh8bMŝvp?u6bTq~W4CH=''35`Cb9!~o-,~I :/OoT9R@i-4nE%Tl@Zqx [VvG5У(P֞[_VJђn< Jwqۃ9gI 7 pq%Z"#n`TʼՍE,5{잽J[U+ڼ|k6KKJzAг|BGC6v{u"&G8q`JU}|.zhlv!y}JM o{;fφRh2'=.?+<.CjwyO~~^%_&7 GKԌĞsggե"d=XVi~ 68oYf,˕j1J)'v:1?ym:l4%h+(Le|:a.?of|&-̖M:S7-2?ルzuzqc.|;9~3ʓx$g޹ڛ 盌v q{Ǫ}MYQT#hئG;zꃡ4+_^_75ݓYݭ6Z0wⲔ5oOETn>^mdU1}rN oէm`yʏ~iD;&FᭃkA-if*_d0Mkf@G>,&nܶzܝX0QH^/;Lm𽖌X9جI&l9djݼr2/xh"j8'D%`YR R)3YX7֏k4c)xLJq49$b0r"IE Ƈԕ.\. smwOGt gT_PU8)AlWЍ0^g4F15L|3ʰ:Q, % ha">#%XҥT1h(bI L2E0Nhd6vz#^>m|</CS[3J2gэCSe4- +G >]I\?;]fv^mo\<*g5 ':"C2w?a5'i/>hj,FPkѫ(,(F6 f(e#pւ{=c ~}>t`c]qIMSXՄiZN4QUO>8{>>NE,c<$s::E0RS$9DDg`l yD&_gE˰I$IbRi!Yt3]ƹ&Nbt lj.`mqZDENYs$pFYɱ0Ey\򛭽ul2Z˿ AxhqZ>븹 x{!p\X1(;tu))zxcLX^/DD$1˒-O@|!tS-`[p(GiQ\4/ :2$LҖpf.Oi>u:yfeX-+Xyh#X c?On89/PQ"cH55$3HhA(¸;ٜ!טID3 uhsp'6JyG9˽Q\bkr{.$9KUG\ 6 _ Aquܲ."^_T .PNʲizWw: ڹ~ hs߰ Wl*Ң1grΨ)RO{s"Pt CHinvo⢧՗_Ӟ0dN-wG}vұ|$IT8Cx9r4SIrb|,xsNŨT")CBi@\)7t?},AIĊ&8g;8 5 _;X6^HL Z40~ b*Yy1C-"zv>1tT8Q5,k AL|kqL0VH2},FBPVn sʍ$%nΦőPCMnh]nyhcшwAwYk:XQ%8bqƄ\j'W|Q+ 9ROkʂ9ZA7Mxvjpʂ.˃e[pO2I{h:֬ࣟ^F!ꕆ8X:&NĻ6uգ #ުz{Q$L"F)VG z&Z(v?t^AE,Znf3T'& :2uO+M)5䬯nJLN^L=^j?W&'Zpi4z.q}ԒdPkdiF[qe舅|qea*3{2n c(4?g_r8CƸ<8t==mER9mt/3kҹM3Xn=%NhNN˕Fz Yq^O/R"u!!dR_N 3X~,P3ƕC_~Z,\ YʹWîG2zmrHg؂:97H_PzuBqCŊ_C%80ƕ_~-k3̱}W&HR^"|ġ$M:uG_+  F#еwi{mՈPHW^b-HJI0U 哧*`hzHLgy k&/:>&΍j],?#8h _nW}z ^0XuZX_4`1tf#8sR,җoGpgZc3f?>uՃ:wFȐ)G̷N]%C\%wAu]!!]Ծ$3mTO  '"p(M1 O/28n0^jc80hnqyώ31j>Bxsa ?v<p#tTa[WV/*Ǻ۩y?_keoZ,w7/C^Kh 90\{^ks >wK*›)7h lսGb?*P>)ypT,6+xU!;~K-.΁_‚~rm8KSGs iʔ";Mٿt Aq==عGcMs7QwZ_]Nrxe?&S/t2s\Oˑ6J_KU ƙ'f1V,W)d.u<{!u6<9K>FNɑ~@ofÙA;Yx4sUVgz1,A{o/?d_D뛼Ŵ4Ao?Tn,}fjɄzB fsPsP~8<_ӿn{}nU_;-ed jy.?jb6SXE5M8;R"z؍p mf8?o|fX]-RnXaofGVf^_j}׬R}x6?w 3z![ZJL>Nج<0b~`mn̙,]h^~kRov|b;VnsuȺEDO6AY<Žt8'xNTR5a̤ƌ9LZ-sLsMM橐)ym;疟|5ߪCqnzL둹O'[YQ^壤!oٵ\tCԷ>b7ud+^}N o\LmY {έEѰ0a^9?Y;]\m浌=u6HY$NXd-UF~P-Lr؁9 F]Fq@yqi 3Zw%B5pȻ>y|=xq_tGuڞ5qys?@Ϫ3t GiY'}(;"eKy8D:`67Pyk~ɲF HC 2[^?ζ uŸ#ߩq5tu t.Ri^hDi66 | 2dEW7>fޓ6cW`}40XI];lo};;#T}?(&x@a)]D?_#:1P5D3G2H0 }<)OUZCZK/yz ~qFꕒh/`|j],#$\!nSD9{)5 9̢Ul l˾æ} pٶWbڑIV.s+0b,w!& QlMS-VݒG'LɎЮ+C]rihmkFH>s *!rGzD V,1JV%p8GDh.݆?+/,gp2Ȓ2^EwtG7iyctڽa(**Ţ;z Oynw{CQ3v EyzQTO<%$=8= {p|ޔC q飳 n@u\]˔LԚi+R`Dkg*Ɂ(0<']<_Yu mJ׹J5|cDt Y qLA= .ϴH#CB"JAt(`zUHYoSHY\l ,15 UPt+ a8mT^‡4dEܺs>%Dum:T\#g]iFwp׹!4.A8d 1);v߯bA@|=T6Bj%EwOr6cFHUkfIW·٬֏pX[1IkM7,<Q!E$$~%7ZN Ѭ,Tg2V>>T!)8ݶ >Xp6Ы69JKggmvzI>1' ͢%/{ɕǥ{[쐳"௓`N~K<̐öD\Q%HʂQݨu5Yk9Z%DV"XiTSKZR "nD8Q9;igx"Z|UwtaVٛU{ZSKk}tnᓅ˽TtS5CX0CHsXϜR0w  |%D! BxDy{RP%}[H(G浱wrnoƗ[`9!|1jQ1ĉ3ϋYh]݅ڽ!X(jgl֚ey^A|!b-MfMB! %%ԘԘ4:L/8[%6zckG9u,_es(<&F &@Zq*FRiEkk6{k!Go7gWg,f|Ϗ;l{w:7O{;DĤp !Ws{pR8y4={>Nyc=zz{B&=ܞLgOʿVYfjrT4L . {6]umn#YI}"8'V`,r6ţd N~ׁBx,/pY@6N2AZ+TD7ox:5b`QH`8z#:\bQ`4mxD1ld,U\3e tf"E$v<|^`l|U7mxUO?9fWs)*tJIdX#-E %Y̱Y,xg 𳬱=ֱ۱G4)O[ . n:ff> "C["D,`󌴄Ɏ>Wyke|-k^4"EڪmoNʫ8պ1E8>(l-6Hږ "B( ^(Au[^RߚEwtA(JitzYRRZROM\+O=%ߐHp&ᘡ).##r>&EIVǟ[@t|:~AnȫsVUmǠv֘W4u^UIꭻq|t 8XMzC0jsdac ma lA+: e#y*%5u ٔKzeX,%Yhea[y /yw!n]BuhLM}n 0`9'lTx^+yXESx * H$)+BpPBbP̐k[:Fr`L(>;#bi#sĢB\PPwHhL28猪/R+ ^FtEÎ`!!To%48x.2*^K1J=*9|v3jUwmxj xzrx-_Ik%'g߼%9rPɉ@ "I̽,҉.j 8k_5|,עbpG1з_js+H1諌*JbC`G̫F0J0p-_rEk'1(6"lz̽U#XibrPi bzo,x7)o/nN8_W;Q?AX[|SQn=GE÷h:IQSn`)/-s `fJ؃L0GF(Uz8Wš:)1EȄ܆Av!% T5޽,9yQ~/^ ъI ̌(Zv0ͽS`c9v$_OPLzh;%m1 EY9E|Wx"}4? 6խe suF:,H0h֡y br0/ %-.hA(*h Eɛz?QjӸX|9~PΙN^s+d"(D*ah@Mmk 8m9 \+7IBtxVXX rq4"NFsbFv:=uvedm?K63Pm?;[kg_Ľηώͷ}SeR)9&6\?_vݳ=z^bHQoI7\Ÿݚ!z#DP#b @C>7RmRp^j%86%GInZ6Pv7W 2Z5}4{PQa>79CJ$.d|J dWAv rJ!\)d˳ԓ;q`}{i:>./O,`rg  IӢ76g^ga縛||'^HZ(m{S=Y=f?o]kDl63`f#i2 6Jߟ/ _~OjguRn.;H2] V|]Xn߷VOVîvɽk E֋X`賙 d9k[_5󗟖z%|(,M9`Uw]Lj Tm |&姕 m %2-潾~i>Z@5fX"Xjqowɬ~ۨ5nQ7u}#2{#8?n륂yls:Z ޿|I1,şu~#1CKx 9F-SF>IDqPц77;Q<1C*&HpM 1A)FXJa)(GrvhXOOwWu<:p/P~)r[M1wk{3fQ7tK |RA^ߟXŞV/&<蟸:tr[Y#Gt- rܳ)qT-->53 yTc>J$mc 0|"vCegȎA)㉸؃'$(a7QhօrQuc9]P}G0AcNreF Q"cCBMfSsuR>UӃՇw{}U-g-Ӛ `1$1'K3&H e0"Nc *}<]ZK>q ۓ"ԭ1jۦ q>OP*yrm-Ip-7?)CKZMN I)Vx^W8cW.fT&~>_b8uQ*pRzT5n乙xxof5J4hk71Ujw;z=G^'bx%HPYg1/CN{.ZCjZ"14!%^ %+)C_Q|z3Fc JJ==<[*I(Z eGVj*IǷ/Q52*rK9 F!ad(E?QFԃc}5w(]G"OɣQӐm|h|U@/mBaL E{̍rO z:1lߘ0{ǿvl if:O75guX|;yvf'VKX/b3Hq&%Hnf+GҖ$\i+es~𖼶1e/5.n[)A]Ks[9v+*o)tHU66I%]6xJ(QMҖT{(ʦH^ KQ^tdpޏ5+nPrEbo㲱T$|rmcV@[W9ÿ9ּ}>kX>b(T2テJt:g`udX{o^%0NQTdC4Æչ9nQ:,R ѽS|"A ˙Xм->䭓 1xƄjJ2d(Cݗ3Y]<8){ |8דrW[^Dz7P2m/8)ȿP|I=jHo?,x/xw9k^7dZSXAM %X'-VKIl ǂOwt{>m_(,/DFhW?,~Gas6GxhX,m~ygƦCC9Ip2hs֦B 6BEG0[yu}:M\$'/E@yp}\dguhDX>[,z3 <͖.3ٳ؈٫OƖ3V1Cկzt94P2ǰo{? ޘ gd ܌v˫\y~kt5s$~_|Z/598ሯ$Yw^UF}w_ʧ},(ipE䟮T}Zvoa>@+.X+!W?;G"`j.Ύ_J4'JSN1:K-_@쾫uW_uEJvH@p+F]4{_¥F>侊hmH]B÷\Z/.CD>0_žoJrFz4IvZ_g+om_T4.Uk{/{O7SR5+ܹkAڌVгpiyAOQO⑎3|[Erw+$cSU*xwɞ]0Թ[XubP[3j+ Pޚ\.~JOMq. F UTB:v9>vCl7̭~{P:G9#0GJR%g]ʆlbޛL/21ߒ4 _dioB@򗙟olmF I?SyHH BHb";CT'UIIotn3XK{J.|kL3>F>|!x|ovU2:AX(2U(\QS]EQ3ɳkΆmq}m)W EPxl@w4ܒj<8G5A|'K}?kMBy^Kنڰ-[g/~]FA\Vu(]/ְ_6V't}QpIyIxҩ,| 5%)h]'ֽK+芉c.{ 7g>rԢ@KcT^t8mkPc1MxڡU>E|>b(>Vw!ZYGҤ@3$pq|zJ"X!6] _?"rZՃX="'ZDmޮ~ !*)˴qF%|ʦ"oÆO"ĥ7;]2Qx}!K9CƇdTWP~j${`18_|: .rf$6p fKtKVpLnwnܚGgpnk*RȗDJip()˨nGgp;:Ɨ+) Pp  2N+jjgn"οTa#sĢ# W!Iu2ʌE;;"р|d!K @}x{O@ڀii1eENKXvN/t |%V `:JM ڠcejEZk8)XB"`y(<)h؈0{+i/{mzV ځ) uE@ 12xto'֘ >v6wo4t(&Dh5uImX2#LB9a#hU/X-W;*] 2bR%Ȁf Nb᳔!Ni)XB<%0S_QD8CvX`LBPXrHk!0<7`oǻɠ)d1*IZ+nGjU_v,yD7 JZ7fMTQc"TbF0J0iT$sb A R6F !&438\6\~-Q D4!'N.0ʨ܍ծytG~Fvʂ<{*()c!P438 BY%"z]*7^C.0~'{^*6) NRޕk 4à @d5CϽ6ɞ5~u3B6nL-g(|9]QRp*2"ev"!9'~R}eh2T>wv~p|Ht ס4`[s:Dą !&}o29f5~E-Ȳb5FG.0b#ʡtIb ̌(ZJY prk0⒊)$e{)S2k<0($J3<]x( ?nxQܺ-sڬEmtϮ^;/u- qlrL*/b )}(F#WD j89[8chRZ\IU؂ؚ%4ƶHQ(4<%gnᡄ_fyI*͢{038 P0{/D;}sLAV+$! ;j97inrk$Rh8t`?>j^/RQ',W): -[zE3 QsCBiE3 7o--.PGQ084O(:! {aN Hq72JJFYnei܀'Y-)Ŧ͍u&Eȯw;JzGT݄:B1UA썀&!9!Rяu i~0׵^;;Ǘ E| 1uǛK,l:wxg8TS/or/ӷj ;ʽd[S.j06xt"C |"@Hd"؛FZ*u0>!Q5!9,cJNEL bbP؊yRoo*) NۯKEwGgpjKU< ͞Fq=4?$, 1Lvkؒ%7qQΨ5HOQ dTIp(~GA1OfZ.b1ǃb_aNe,+(\ՑڡP7%R9ۏ3<BJh|gH9TĒZ1U'h<8|E$EQd@gx7<:#OQ:_cbyjT-^_Y<:b~)qlbz &KpUObݍZ&m7˴rUVQھ^)ڑ638xp34/C7YKcMsQtpg$^qUAiPs)Xi0d /ز,7%പ=o܉˷َ#{8]YMkqi07Zt7`iVU232 C ya`}--}4+L}!ȟVpx|/M+-Nҡn_zjJTJ\jhi;Vv!N.=Fp~sтߋ.Di/s}"-@53c ~'"TR}2/8\> 6¿QbgҠ[}yO? ^SCJ=MC ˨4I?AX,(' uR#JUX` 8}&cGY7 aqqȒ}>wEBx vwDX8yS1(:38J$,hUQLL ɡJ2^.jh6 CdJQ/tF&s 5)p[\ȼ+E2ھ['q238=8,x"RfPgm)Qd-_  j؂Eh@ . Ĉ$Ka-c?^4fLɓ\+ j!CEdA4l>*N׬: %C^RWAP6P|x_Dx _b^ƹL(]̙l8eJn>xͣ78)byex,ޒ .8 9>fѣ-OP9axg8;l#I Qc)̍F$\D1g>@5<:?(.?z-#'2n 4ny#~K7ފth1-g4Z<;yF|>w3f-~ݵ*zyrSSl7#cnpb{g}UNje~8j86(Ax'?$]S>W_Oh:qcs*VzǪX Z˸(;9,DgEň'<:#IG4]OFBkVRd"bȌPUeMNh6 426~+M TZ팂'D;/>zqYQ5j$Y$fpE JWCajtͣ78vi+I2-ڝ8v1ARSc2+u⻨(( N;>5E4, s/Byu~*Ǡ7,[&?* @^ Ej7{rHBQkGg|.+HM1`gSeYѢjONWe$u$G.Ȅ)@SOPRXâ$%bR|1'KJiq-6Kne*E=<:餾 VjѰ_dj4/`gZx6_!n g`]c7;INhv7=XfWEJ13(#60U]UUUeon19x ៰yMj):Nƃq\(5;òXx3Z:RBLceLcq+S' !: &Ryo+Ϟ 66üް|I4Zص{A2,vI΂*XV ^gsao{ַ}Ex~Ljscy:#7k\Hpr(BZiO>@V]\_4׆.n]'=4m(Tp\Ӣlh >"sނZ)ڨ~X!l}R,_'sV%¹^ȣsWcM`Owb1qtO ^(f؍F%ų'ݸm| : 'х&XUIJӊ3EpN$Y CL+rMv`4ձ?zj݄ s"…+D7-W70#L~G_MݸھY`?k!/mo.Iށ&S!'oOחuR7ޟyQe  E-&uHm} <~,FONXJAY8?Ktbj aKٍl&ꝓJI!lۻ1:3xrR8`:/b3 ֫YSJ*s|L*-MWCˮ /KBI(N$i/ƍaT HXƀ!vCOW,}TI5XY15fӢb)s)xe":>8saE+ü?NR> QhBdj-LPZgEQI""[EO؛0=tC?yÿ* RWBe b@ 4h&El0t&u 1`c 05`j6lL͚axOW2J# tc"rf U 6:d]Z*j>6_HɄ׹>uwt#lcq:tYjhk oՀal H?3 {Us8M(!ϸIVh b=w$ ;ߺ-}ЉHiqǚPDF]=APgO]+NeD\p7 SB7oB"0 q:D+ ;f)ua!B JQ "u%?0PÊ,QThx5C\345C\34hjPii4:X!"֫ a^ & F gc<6+wJJ|P3X1e\1e沌y'uVS,J XHcHǢW[y;u$7j 8d6 J26|r٨"5r> 'W0'WUR+B⦘TdΤ1dk߻VNKbRw5,8Cohh7I:ؽq٪S],fyF:F*TxbLf6͙d,Ѥ͆6"T5q ,"m:^7܆奋sRڜuɀ#v*4RZ*QP6[!z^WYY|H6"hWR >:3?ɏ'oY3f5$SFE⛳:'+r œxRӨ[̚^CZ,L$dt) Yu,/0'a:1Fo~_+XM(™<.I=@-: T*%Ofs؝w2$5;zPa.Xϕ9oldz]6hI}F NBj7X7Z&ܖm5h"ҢJTB !H(owc0⡿٠v'N\p8.q{yAʩ8oq~s= Cа6*~0#wxyv9Ǡ[jG!h4mCx I9",(:C`:zCZjYc՘?$Tv/li:P'e/ӽ6M(Srɷ{ kҽ;;Dג{&_zUMGrO1- 1r8zi{⋻NmiFϚަٺ[m]K~ KtOXXs"#XTA=ÜYTkNerC_d@EO_Ģ5/n%aD7U.Z~l[R?(}ف.Yu2m뙷*fi10.J1m\qaW(nąbHyFʈ;k'ۿ kj;qI){X!T_C}UO4D\! \! \!J!B=BvO[:\!~/`\+`XIcτ[+R}DBfNaEO]MO[:\ܚp•[qBp eN]N[cC+),m| W_\)E?ݿ|?L.?ӿӟ|ާ ~oq@x@c@jzP˺p+Zx cpڜFp场Lxo6L)"ӔQsiDNW26NEp1 Y!樒ju$*ZXklq5"8M&ܶsV6BY/zDz!t*cRPXXn\T1swfֵk'1etQm=j)V[OzڬFyР k1E}%0i 35  3.'A b,w1f֤[VQ}H^YJϵwDQ{6W߁O~m)d8._[$4uy+Oёf?KXb4&R㲍3ތ)+b_W⌤i?ŅcR!vCrK1`  γ3k"&ύ Vf 2-S1ɒ 2ot{p)\[Ēu"kY]F|r"",&GBE^kW$nNV!W80͏ƓqSU1U;kҪ __+:׍&U_;1u_ u1X<9gM^m8o_u W^,9rlؼh˚cv1Tm͊ϖll,'/Qh,/9e_ڷZsh/g2+QR`ChvIh>|é S jerzXj)ѶIA>xWĤLh_Tp/`򤛯k wƬũLnr`1,۴:e4@l8ٻ6lW=rc טܷAcT)J>Iqz,ugNS_$jnRD]*ȝ*lUp'3h`Bc'4@VniPBŐ (zF%(!e/TKTr.'b !, 24xjYsL$$#ETjy5˾609VZ+}*G6Lどc?07`[0C5sjYfx (:o;c-8:0{-#cR񥐖c?W4ZS uw+r27, g_ZĻoMos~11ziK^{MlpxvVh~9^lOԃ]1",DDꥦ0+Cʘtg+8ݗ;fWG3A݃ϚwI9,Rg2DOeLTHJҶ$J7dbi1g;Ja8҅0,g.I|fDV L AaEd7|{uֲy_x)$C 5Fx/Gxᵺ"WyEh:=V)n[!4&na[̄aFhNP`9VM̧̳5goB,ƪKBn Sʖ S;( 2Fg{\nƪۜeZ)XϪG:\f`k.PHx3jWsWL#,8ZI}q玫w{V' VWH$;N+sOeW;m qʩ8#n43: # |HUԹ(gCo[^p~ڋ6ni09}r 铥[(Mz{ɕ\wYq0?_-v[A s?ilU}Zӫsh5Z`(3*h2:ڀmT 5!Ӆ/1^7IOVp٨5#!I$:Xp}CmؒsHvY‰ .S&xs^}A ΃qtbRW?uE $Kwi<5}5we^ևyr?-ZOR'W얚mz{޴9D?GXHzxޛbsswޢoœCG9DŽRn&x9|YEw9lVJHH! YEJn Zvsv=׃|s+]~zGzZqpa\|IQA)JX#Ɍ.jϼ M?u:||RƂZ!v~ 3;ifuP;⅂+ʖwq%b\*CA@g OeI'$? mx!aa j[mQ<:6~;m{L%1U#&Hq`|_Xwt fuD$YVĆ zdzyPT3SގN:X}zy?hWdrgm:v'/'}L5+s*w K1$ozq8 f,4b HE#GF\/~zqo{KfP/ fb&VHi- t@,b+"a!B{9] D9U#g3 ]Y%ol{Ӷ p|xi.̳G/} YtkC_3:ʌ9Z 9i bR|>2<RS5m44~(LI7*}z! ݁Sp&85phG>\G X' WWO?hBɆFSXƠ0-=Zj97)G7|z\HS?LCۦoufa|7xնvOTCk:4N`&w{wmV}_O&p_D{v){ݗӡ.\ϗz=T_a|#fPn:Qc Y0&Ua , 9o~Ѧ𹬹)ӱjwQDgagdW7dҍD Fjj{YmבZvY{N)Z`B<UX-zg՘IDk1hFs+%=kgSG_M-g;H -, Q;7 qfA%*Ű[)r]J7vޙ|W ghgu}$q J]n{D@VTm<.;h-R܀+! N9ťoW12#O3A ;ѵKqYv(Ǝbf]d8)]mX6|7 38ۀ&hd0'ZK(Iv yI"R:" f;iXBg `*8S:$( r|-(!\0Gd1]hس%cY.HǴ (QHLd$&E:łKsg H͊:b2kZִiԴְ(ņ=>Dmb󤫎fѷHB#")D,9YUNyDy3Qy&*τ˙<gz3QyCz3QyVLT3Qy&*D*3gLT3Qy&*D器<g򼀒s!gh9]P;:D+k+@A5AD;5躌\9bq蔖`ٟ}JΕfoLڽloI@ڤvيtyXCO4sH~L+isR{VV>gZL+i3|ϴV>C BTby3|LƜɘ3s&cdLƼ."h!$!fcaVQ[A8*rD"Q#Fc|>p>Jv39yivGHv'ӠeLI3uvd/e˄:;kZִL3uv~3;/70}WMv&/E5/](8L0+ vP{!;"BxQ0|%gS4a&Sj/8A;'Y C_Ce7&_%Cj=`k W逰H ~sVk/;^sXewQfs)-L岤ִbm&3|mm FT^3썡]r_P绮{O$v>X87hJt>粬ü iwr0ÍT Mj5/jZFjm\iTGTqtx~%$Ҩh! #*tԬxOC*V ,Bռ=DGqouj;dHKB؋am%$*c"DNV9yTy :_}u>l솮[/}-t> r4$ 0拪ańMRi!ve+gƹv>qgZr4ZfVԳ|Cބp3OM9yIS-ٖ(W 2K#)6gLqZ' C!R\Q}* z!fH,(Ā. 4I{s~;*ZMB`)II% C. B Jg&Di~nɱbWN%EǮ[_džә :Q] + )BIA Xj!jQsz!j,+EqJ$ Q&jID)T08`XVGC8;+iU-)!n!:f[aFhNP`p:F gk|n gg>s;{{}ŧT?|oT|G_>%27\_q~?~lorx0G&oYpLKIj/<7)G̮7zk\H^SALCۦquۦ|7v ߫mCk힞}csֿt7ɋMXIV}_Of4zfm'fH{ }9]9z\Z׫?x'sy4t07 `A47pQyWc`jqm`Q'\"ͼ-R[0kT aEΠFGY{8OTRgﯮ7!{|K[i | r"{vںɥ{6K]u/2ݓ"vֻ{oJv4ϟ. x2~Pᇅៗp,Z-bfmZ_9}y)42#WHݛ/KVF™l>Gf:փZso-jPdd=ϛ3gk!/#_ZHQb-h6Yo+fgd7V쬝=VZ>T?~k;=vCaJ2r@g-IMƪ36]B&x,ə=v챻QAg'3viκ|w0?o>.׶kP(i-uy[,rk?_6K?F41+7TeQ6^EFsp6)))'{J3(t)r^U&Y"p5UgACDNQh#yFBFBUk"tzr7a-Ӎ1׽ASEow3)Vd)UeSTd c:;YIoe'F╲@c |EJIS-A1Ҙ lT50؂ gڂg`mnhF ۚЮ)]i72Krww'ߕI9ZS쏟〼TfSQ$#AX%>[J<)DϞODںJRXF:dKr6Ic+}Ċu8i>5|kxѕ9_!am_~oOM\lceu1zV TYKπpZr,j҇cwޞɊ|jCկ<\/0v^*s =7x`v\n;ش\c~/L|; BAFCBVy"jGEu%BH *m d|(YVcZVIE(U?Y'MՇ-럮+ڞ=ms+hq i~\;wΊvn8Kpj zCڻ=FC**o}@\!"K_ X R>[Ir)Z j rEVhw`Z´^퐰%?Hm8*ZE(b !lP355&1;&h*bvNՈ`elc1Fb15;9UnxkkYYv=4zUD8? E%<$ U~|YG;Fe|Ip)zE7 O$j@'9 C{Mr 9J{[+ЦRmཡѡ6/8 6o&+!1i`uv6#a+9mvJ.I,5P#ax>P#$NHcj$;lfjJ6N0VLlڛ26\7e2mrΣ$:|e2E(2@ m/E`L6h|̜YC1kSk, K̨(JZch9$D%Q噋RDi  6 ȵ}O djeMiI=]pF%ڨɡ\% #Qu^ kh[EQ!a,*KIFm&^YpJPh`DLEh+^$yb)Gn4xT re'>/f+V~3:!I ǜVVrN:**9B[VKd'y?^\Mcs络-s݇y oۿ5ֱ3?I3e6>z}#,~U!jG*`IzdIq[Q ۨu#̛>~S^/|bhY_N~@)>`4ۛj65lWkQ>vهj͇:]OýqWPMq9V{GiP讐Z=틻Z;\(Ww rW jR97zXqQ8؊bj/خ^XDՃz>`+h y_ԛr1'W!# 0K9Y3'2+l[.4j()j1E][oɱ+_;R/x7ݳXoB_-PA^DR")KL._UWWIOo ML u:E3:rsgs @xO[4̂04̳UZtɸ&40-V}wbNՎK/g,DJZꔵ̹q& ` I(ZY_:INK-'-DPpq!]>IXò^ZK\хkтŤm[e +4m'%w9B=psCo Q;Ed< Fﴖ?8Ǽv>sG yE;T]R\~m_Sjo,5ۛyS(ӎONLYtR+e< ,M(bl *XXB[WώnZv,@껹Do57eK>be7 Hp|Lah\#b(,Y6,r%໿k_?ҟz)q1BKw뤅v=403|L )+48*>#R9G*5%|-4|[ x{D%A=&m]p Z}wŸ`xt}R՘0hz``贊6x >ˌnl- _S3Ou[@59=v_G3ydSe9d͵&:& &,xBJvD Bx>cW"eNԖ 2eV1'Xr'j"Q94bk聧{W@0.׻o_,=.=_XĥѠhav2]=l6Z*݌.m&:o}Na]&կnH.r׼;w-#oom`*@:˨tZP(=OQH8 Y$F98X*70,ZG]™depfr4njɠЗs anl:9x%z|< F\MNl?eX2uFlqL <0EĒ[3AMlj 24ey RP18Mm+lTnS5*_"!;u >)l!rCi GcM:1D0+xk"&DFE+=118Bhf ^>BP2 Ο?h}+msFZ3BGچBON&!r@pU+Iad-myzvOּ:{NLШd(: t KͳS$XoYT>qNiR2Pݨko0ZH/Q}$Vа sH2ʩp nFP|VG߃/.~j. b#2Z@mT-:::B)-pO͔'Kpxi=5` o)&⒑)ao~Єj:K[uV+pNEa4qEqsBKe^oـ&8Z=QKhRD;D83|a_=dwb72]@J3r )4qfT2o#;16_i*joS+ kKy.cVs wFeD\!(sBs֒Hc^#x3bQWUg"cN06ף>i+ުzJflLN3R&BlTp!Dx}>pi DTٺ6B " ZSeDNpolg<2-ፔƦeȌΏ%^g *fOooƵC,&o] 67t^Z3܀uK3T\D$'2%IC2ёh,#=%twǤ:xOzJ:3?g>&;3@ю4z'):/Uؔ0,9O`ޅ١v!5zaο^=tLTާ }yzPjz%s=`%fgԀ9) N59甇3WԳQ$9`9Q&U.:x W;Xnԝ@:JNRq:?~buK;:׮%U}34_BiB(*tjv>gaF0ϸ$Fك 7+.q|xq*cg&;fՍ7ɅwdPfn?.+9?݌T$.]#I3Z;G #5̓`ØJF *&{wӅ.^uG'_uu=VTSR i!8`WG` $ŪNk:U@wowBOw߿~x{Ͽ.(3u+0JSMan5:_w;L6547;b\I|qlʠ톙f󏝟neiN:..؉ӯͯiٻegzߖge%*V"x q90qw iK<_ެĕ`N) Y6s~6N&$be$peu]y`pWmu>fFe-ߴ|s|ro)JeedߖvA*ւ뻙Tn-&ni߿iRv;,'C"ft@;R әBx*:q&pa '>J½U:i^vy,MLw@Ua"uo*(wzr*t鋚L˱l_! %վ0vf èUbLR,_UlKK-ͪg__w31o&:Ewby:?6v7k7W)_^^ywe @$*̔) # G)XSc[ N>Z]Nl_%3[e&sJk4P"2}s6b1 VҤFȘHpZM>28d}*-NjGvLHfK1ҥ~́$#C6m;o|& 46Fkn,<{1vqM{oud'LW^s3> M>-~ jy?p`vo{NXx -f9}۶Jm߽$ 8; O*,֕xWXasel?"MPRrVIcniB K@V(ĴxE0eZ~;W:9CckR+ I[NsKˢ>IJ2Lx#4pz[-yts]j,]saԛS Zp _*TPSb !6#a)I{/HV.d̴ej՝1݈%Je[f]dRA(v-<4ݝ `N"kE'QZ5Q[DM]xO%wT$uץeV`*&Xdj" zۓd -\ Ot-8r@܁uEp` J˭4Ԭ6My L0-ǵ㜣IgS)3nk( L.)1z$}S-dwxZ4\N)G!9X-st=R ,BxP"p@F zܶ)5h5osJj2jji]Ã4u6^2mXg_wu`s F8Biⷥ4ԺʴA6!e^( 3й5ܗ-Ngp W?y7nUpu#6a]+pWM ɋКֈщ괍(mU#T3T sS )pZaN.Ch[ +ጡ4]e%i fh"F]*+qSU֜<( V-zhb~î\yS@K19uvTeW/]q̔ bWsc fpn hdήJ*Mˮ^ X $UX6]epj y ]Dv/鹱rl>z2R4?~ӏTE^/߼Җ$Ec8W(Š3_u*޼ S00Kq NSZ+|1JneP $ Yșe#=r]: g00! +TMpcFpJ xϯ~,}^Ÿ_>~>4o~\}WO E'#d'ȄheTX/֜ ͏P`ޓ`dR1DǒE#=K8##JxMOjA:Z~*5C- m!ۢMECECP0Gc%Pjk,v0:_Mu :s%09 F!* V`vBm;ɶXi.]f$KJ>)$R!CtzNێmJ*w_K ?vnnúZ]7ח{Yao- {Dv u)pl`lAiHʛ\B^pgoHN<2B1ȸo5EThR /s)$m\8SZ=qY*̯5V^)&SֲJJ>/"tALtqt+`𓓻kgq(a#HƨR-{F "{`w iM>jN1U)ܝDwR16AR͸`s8ϸ"S`Q^$c=04RkHKDh*"6FB))gNϝV;D!RaK!)x!3S (G ZǠ(2 voAZZ>_zxlelmW,L` QjO`N()(Icќiˣ`Ҷ}MuwOmZ/@@`%p!8%#<SpDAҤBKHRUAR?b]Z'=gYg>&=ܙB\IdE n oULqt~HY%ߔa<>:{aƝL~^wLή ~Y`ׁ0Y1'~]Yh07SXaz6gJ2F?qr536WK-J`9 &Ut6):׃03pH.>gu3=GE]ɕ oo/-*g\9^o~u/.''dEhnK۟m/p+vRgڝXU6jfU>< N0b~nzݫu ܕͻNjתgU9&s<4N)$o,}6W99#$`ŊE),O.w|wx û?|߿>~&T~w?ΟQU , I(mӻ߶ޚaRְtk[?W \#W{|~fZg >u~w>{7Ҿ"Xcs Vq7+]]C'L`_MbCqǗKPm*#؜#P nj@_C_UTW袍FZ%Weee RpNhfD2fa*0C⎪}+c#Ēcoɴ]2akMbNyN{<^ 9Qtئ'=IϷfSRַsYer{X;W{DEf+5F0L3 3M t˫{vޭ+q]]' Dm t)}u{O[R7_O ftIƦ<2d{K ߩu z.[ZAZpe (UTnpa,-]. 5fK#˺ :?2aDq`XieTJ:N(Zu75I:x\K4/rV\-[Gec =|W؏#):^ 8}};WX qo'h75ߓN`8~-Il,o3HE۟TPy=kmbpO؎xsVx&z\,cmh* %;J :Qvj_j_'7 ցlqq87S}Ԓ y#qh[ dtm(sbUSGb){c E޵u#ٿ"s:"  AkHn9g1}owK-Y[)KvÖ}v]9d<)Q~ (gT)%.D^4^DE~aTh5$?h;?fԙOe4%Q6 h/VEMEwdxam4ƬK"й]i(iclԢsqiߨy":Bͤkسm4fa ^\C/.Q}OM:w纮dn>CB1b:Q?y )k#'4tbZ9:=DЊDPLTfza/nEWKniIo{S3^_mm Guʹ{z{rRwRбVmSɒRDUhiۊ$Z.yN[?4YǘiFwS⺨ EUG%>C[RiQIPJ[jh1WׂNJu7l!@&W y f?/s{]*[ˌ0?yN|&3:3in/փO^LEQz[oyVk4۴ںL.lm Kά)m9J FtgO}/_A64l VBJ EFRR @( T g%FµxhH@j42302g)O{8Nqr;uzi*CUK9crmcM|Wx68Q dw:h ,";Zڭѐlz~[ù=wi{8!&>ImĀ !N"ZUiͦ &rx-r D4j$to57m][mBfozu~{{R&@O~PGwR<>{u}/g;uׇdǷ=̟)XXڥCpuGp[躞.R= >\Uov![o̿1*6^5-K[7v43o_W*yxge073/?q qnYI; /b+ Ee@gSf%'LjCLDIB" I޻-B7^'`׮S ~ȗ6㽫̋1a뗦O;帣_oqҮ/vݣmtx<]5O?|wor}dt"kṹ>=y>r5tͯz&Uw_u{sraLNe( %6ƙCG.t[&q޻a:{u}WGѪ;iy=&SJ?8;k9^J]^rX]-wu{'?;s{./9?xr.>6/׆nΖmv i7⶧n 2Y|FmtzvTG;Vo#h8|@xV[iQ9d;?7NU Xk\h<q}JU'u g-V ;g=.Oy#T}m/:[.OK}k.7MtV*KZbMRgM`J(gIhbpt׃?$ICM%[(6ǒR1Xӏ{EScUX=5NX<~oz] GZ;b$VkѴh龎~~ΗZs.d֊*L&0)W*bDLµ,]{Pwftc3kx38fsGMٴVҴ7eʪrQ Gp5"Z%2Ⱥ0dҩhƳ{klY*K›Lw(#FU@ȿ#0})Wt4Y)n96bg.+tQ!(9zx!?„#.H_eK8?\? $kœ5c'dL|Hb'토D7PۗZl=r#"BkADrYCیLt0 )䘐@Ot0T2'Ӡl}n`8e/CI z ր̯\( eC}VlU4\b7_9edǤ0k&GE!V ˎN h5!]*_ພ슅@Q ){ T ԫBwήjqbcXK짩"Z:iTN~iHQZ\ <"Q pTi]VO}Md~Xa}V{~t>9BK̵k+w7{٥KGPq.g+knHcd݇"08lk7cN(J߬A"AHtkl!OZi%Xu%T][g*ud#s\G:2בud#s\G:2בud#s\G:2בud#s\G:2בud#s\G:2בud#s\GeA{NL h1bEMφ e3\"%21ZL d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&e ]91h|@/@\j $U3% %2$L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&aKWt5RwWo%մNO!8w?9žqI}>%hgC\JKJGt1#~F~-)~*-t9O)Lb+{лNq/C&VD_A4KTYDU@(AvWoo?A^w` n=<0/+j9/3Dx=AHXAn-++,,帯^7.( &R(fa1frau5W7zO8~ew4JwjJ.:ζuwR(do ZGE֣RU;!7ob/(*T? EvbES[֋mp۟%m,k"{J l#kkbh1R,a ƬWe5έdVi)Bp K hIܳz=ymSl-3q^G-!EF0y(^9(_^y`ҥՓ/-|CG,ҊMA[ʹ:Q4Xwi4xnWKQU8Q4Pq6)O̞KSi$O=婴R Lyz)O;uqa\ǡ[KtSHT-2 )Tk^ߎ_lca?.-7yE3!>lpְMoۄԧ?֐9tIQ{E6j!RDžHZ…+l[`<~:Q>vA ǁjkR(PatZ)ӂjP!T;tiIgdJ;sU+ZUi!h^b%\#}RO{g>-׃ Lsfin}ݎvFDS|r0,N Z^7SL ]4n=t!6#΅fF3ՃC|boBLjV ЏtTcp[PFGLr9U 'I>,fQ1̖6upOʫ{m|Qp3 KBs%QbrCJi0~!B-M}=YJ 0_ Ϭ 8oŗTD Ory!Fr%M7ޝ K-S Adq/ؖS?i?x3z]Gp ;ov0fyfsqκito٦_ʏuW\\[Fj2r@[D"(Θ'})),8I\u:ϯ' jSQ_!s,w[>}鯓h~ͫ8ݝ7E'\-CZv RuJv˗g&pFlC{>%,0%K Vj5ll| l? ub a>c_:׹y+N;F}Y>iDxX+訜ZGGV~Cr݇Cߏ>-Bx|ȉ aiamhmwrչ^z;ެJS2ȻP f}n9w#Q8.C6m57̵385y@涩|w1gMӪ,<m ~ӚGkB9s9hRdl:c9E&Xۊ3ٓ/7-U ۣYA+|#7LNN+׃qIJ򪅨&jDbJ8+$RmwR;~:7o/?7@uA{؊3j+[dXVw$AЮnGfr6|ztMY$g)79޳{iEɒ 6Â̑ rZE",h4o{ ֒]YSt"GAZ%tzS8W~w?_ [@Y̲I΁.]2O2V uL %?HOЏ[LAo Łї ? ( ٮNk{L\\JJffڔ ($F39RVگs-B¼WO ͡]欒C.Xqq+"IRDUp>fI*iа{˗ccHǬ(-ZCͩJS`(>TRTdB ΄Pk(~б-iiiԴְ(ŢgA 4I׵Uv]ʎ݄* +a)FS"UtS*Ϧ{i.?CJk哯^Z%Vtݹֳ澺4\JTX*cdDIO+YZKRޞw6pUB<ᔶCw԰"vlԒtթh y<癞`d[[ZY ŒɁR)!6*fQ"<>X4 &*l]F[& ^" ZSeDx) 3xeR11i5'>jKs$[kN?/wȒ%Pk4eWr~y5XW R&~or˴ۺxd4 6f LINdJ<8!!d#x,u%O˗d>Bf\ X\JLks0]7T82XSP6hDrIqY4@ ғhC(h&_*늲6qsfACh2r%"e5ա*Wyj7N$ɕC/1)Dh{b΅LK*,d NX`r#*RtH`ks)Ȥl7-^?5&091vM-Lwϕ&>M:z*4J`K:_٢wCxBuYw~yue~^쪏z(dS8:kdY,qA e50x'bH.k9>_'`/RVk"ɕT uR%RO#YB/ d_ Y;6 ,S UbL4IɖO̐")IIÛL3.]zQ1麋|EG?) blu~u2pԫ+fmB tJr>_m>}E9Rl>݀J={(U&,\/_|Ç Lۿ{:I4Man5:ߞ5'`y)m Mm1|ӂo1.Gm>røg6nvLDo|c#驪JguLЊ t%v? L uǝirVX#k_i0.\D-tE,,˘-bcvN3JLP.Ec쯑YYT`#)EX?2㫟a㷭Y!K#Ry[+tΥ^Do$QZ%'*nKRoȔ?%=& ב1OLR0Zy,=m)?RmE$Lh q ǩւe'T+Nq ek{)G(POuEՑWuw_JWw8{xcCx0$<A^mY;R9B8$E"Lk^p Ddso4I~Q9[X;qDebKUzK2{p1.m g43NS^9^~_'}j}z]UUnan-d|?{ʦrpnF1/t(fgE&S[d{&e4$?ۿsWBw{+I Z瑩0#[te}e3eL0GKu:?0ae3Q= DSÖVz!v(4GԶasn 4LI$mqFx>IrLӭ:1˚-vcHJ,ʪe xi\).53\ͽCxCmv 0uok[-3;Rq"Ӱ@u{CNYgoGu)*e\mULg 4e;B9a&:^ ݦ 󋍲Eq8x[2`C]P;:D+k+́5AD;5к\9bqh5}g.0E)JiH* WuJ%M!? Ρ |!d/\%{ c.Ј %r}B~$OY]š?e.O ^Q"a`bCw"eDbYRbo9'xUJ9DŽRn&x92J#"( -J))$mt6W>τ}I![c)Ծ!ߟ騧zs ۺD+GLՙpqvÞB2PMW?Q? E"Pn,k'1"4 @3=RʤVdD!qRPRaI-GK%Jg&D*O'^|Eaug.5v=S)-#8%\zic QqO9DX@GĜG#/ uթjTwLĬFPPCӧB-Yw=g"*|Q%8Ug ո0i(@QЈf5G`1R VZ@-Q:  l5vVh08CX(7ڂv̸ ÌМᴌn[6KTYg)GA>` Q{d؎z.{dz *j-W/we$uv:9炁+&a/fAY[DH(򅖜N*nAfrFU`D&6l-T݈R1T)P(^iGjd^ wE%K>ISPB2rUL!X~_𿛞/ 1|| K3`WaK>]Mz7D`R! 0o'pvUlC͡E5bNۈ  *>~ WIW`i5)7CY|wV5ys \yon",E.Lav>_@2gMHwϚ-xBg͒0(g&Q@D9`%ba:3T?%8{]j`JH~@X TUZ7]`m+U-lYJpw/yApvKuRu_w~\v/jz3,oo&7ba_~T~zwk mk:nMu8vSZT^yOWтnDTU-DBG͊>Yˎ8RGE+XxF9cLF4H*-T*2 Del\=UGB<UX-`!띱Vc&ye4zl5HDhm:[U%'Rm|Xe1 w3Uӷdzp7Is6,Lo-f7t]K 'euK1/$NqfW M6Ư`ŤO35^ZUsWmNq5dMڸݮ[=7n#۴繖h0X\Wbޮ=~co@E8S):ۚǟ6ݶlsmwoM7mq^s˦Ø9Gf;zaꙅ7TT"Usf1KQ\X0taE$TTZUmXmNek`k.PHx3jg>MLj9+f3];8oO5ˤ=vKc}>_%0tX,Nz[H 0 1 (VSS FKke|T 0FUt9iYN7Rxϓd7Mv`m"!z!S'b2(LyCXN%`-q!+9A#JVa*B Vhb|an&WuIm^o]`h;PԜyniPG e!#aF%(N1f}vhG`(3*h2:ڀmT 5@T!26g:N,wukЬWd@%/%O(&L);x4]N%9VcGN9[Gi}2t]z?ϿO}woΔdT H)lu3ߜ!5QD0>Qq 1!1= d<o3cIZx`8[>6" +i<N?gc^oߝ^ /ỳwfbu4 #NňSM L2L/^ׂzJ`RGTPJmrvBTDc2/*mq#~uQm->W.UmC]OĪ ۬F1=x/\Y1_N(3GၱX1 )qxm=mE =)FlaW~㧢 A% A IRHrIi6Pqaj`΁Eaeڔ6fOsG٠]f}Xas|lu<|p@+vVin}շ[Crm 7-^0Da5< Ў"wfzmI[cծO}]gVбĬzXT&n7}쵲2aDD& &HSE57V"FM-3ꄷǫ9ɉo#A5o@:@U0"5 /|vmA ZPHC K)FDpؾ8N! K*HRҙYp@"Ł i#7۵b"{&CD4"SDc(DpCZ,-,xGI4 GZ6xxhp^]P[g`:헐eƜ%.S}z߈|L>~;Ղz؞9'eAYaF70郎@$2LGg}"d)u~ɆGb67of} 3J)[gw^j`E(ּ0B`Dd7 I(L"h421%#yIf.Ֆ%>ɕq ߞ-r-ہ͸S:;m>c5xj콒+jnEu32d9Jٙ1]@ ~`I-ߎ;X!Mt"-h\0#[rrtYL+(3f&f)LX(LL`jJ6N&(6li1ɖm<.tap0\ܒ wN3Fx> 1ni)^s%wjdbo@rR!2!ՉKYJ뇼< Wa'q:mxmsVI4{yePH^1 HVqP I$b岙VD[FTP$@&S <ʍf\rc2z ]W芜" '0@^r'y]u|Bծxp:zpxpFM4h-BpҠr-%Iu9݃NE!Ih3ԩfTFH mYT=L0RyDp;Jp Tך3r6kƻQʳ8^ڱ..| E,Z7@yn{ӫih>u O+\&bJ**F(7A[A+ebJSɨ0LÓipgYmrB$f6161aAX$ڃ95q+Z3 4BU% H*xEІBQ D A1`&(TxyHڻbfޒeu@C&\^xԂ,H>`(PTs0!Ąm|{ozI%ǔ-a HJEm_g#Vjjm+=SaJERsq#4_ '1m_ث~e&G{ok7}?K83'W#زG'%]& r5!O80jy\je\\T)‘嵬rKT\D$''hbS%(g@X5I+0Ms5'U L+`ZKȀ:WԆ 'Bk4% (.h6q_*C!j{߮n2qsfhLܨdD: 8yeFpH\ӫQmw\\ȡAQcȅL r4) RtZr(wjOg?/\lhga7[fОˎ'rS[ &sj6 gQ ü 3-Stߩ6J`{se&Ra1;?b|\ gpU\J43{4EJDЩY.i{`|Kb4_<~Zq{&K(U'jZu?_;.O VMbPjɳ`zf]F2YV䣓Wn!!-#iF[GۆamHaM00h\(ǔW{u g. m$}IAhn ~zhNhWCx[ 76[+I2MAY[XY+C`r|'dt֠#;"A<m2yƒqThD,:v{Ck/:*k;_~ӦW}f(,nV.~I䝤M }e 癙#/=/[umoWu6oy857z^6畖wd<Ź߂:bpM<7tAM-Gv<LGpGm`*Qfs3DbKvlwFΖĻP^LƟ 1 u2XG[R<>a'QF(L}LJD &"BR&:bE*(4$idZa>4G҄LIPF<՞J1pPu, Ě6nt`J4Zx+u4},*O_RY-L+:]+=_@]LԐ U,Gr(V$ID4DMDQ b6g2Qc9FeV.Ý{fA -Yy69f#R1p 6Pg<;=ZBFr)+ ct9mtq A繖IJrnGK~Iis0el@v:adȮ:>rxvdSKp͚0O+iKeN)i˓xN:@2ȕ+K""[zm% `z4u䖴22^B%!:+y!@Fe:%*E*gSMoq_AI?}MzJF#s{?okw%㥙l-V=S6pڸ3\>c" uJ.,:w,~rgw,e)ʃ2FZ I{$d:# a}&B7۠dǣyuՒwiea m Ůw go/؏<5Ff_ '1m*&_ث~sx嵛%^mՃ ~;F~?{jr: ݛFm]@%suSdmYZخ륲ܟzo"a{'W5]$G3:* "!^X0 a)$lli;xg׉>#+_Ʈm\ޮ)4M6]uVny-9WM>ur#y"8E< cAZ]P , SPK&H:|p۾jI6~Y*#R"%1< 3âҌ;$$" (O }W+J093M3A' º / JJr*5b!(tVhcbT8cy3t?ִ]=;xp l% 3ia #(UpS$v) >;\%]b>2CGg[ |3u@{ܳW-DO`%'_򉜅Iot]弆tI8Y>8ƅo$/LO]Ks>E@_2ya<اޒ9<`eJe'X9SKtr&SnN 0Z(rL`MHKu&"*n7 }w:&9M|D>4=Ws]JLW Vm|tʛ7Iӏ &0)I*Kzⶹ &lF`r$}q sBt*aSoቦwۓ~uW>T]MΧgwL^a땥zs"`W ovoB-3I[gb|LmӐi8ӪuyVbpT(Xi' 7B9Iq^'m66k۳Jٌt Is`. }UOpoT@,ijPBؚAĿr/ρOO݇ӿߝ?)&?ޟ~v LKm$̭CIQHL"kj*ԏzm ż|y ھY+S端~R˔+SCٞYwH 8 }7_A/ CTTs*L# _a0/}:kIY ȼFZO*F1Jo4)o?@u vwOƾy4H)ʝVgؔ?F'ƗmY҈T)K%IV Ic K:#~.*Q<$| ,6$3 sWriGһKa2DA2\j&Xw Gou_H6Xp"Iv"wNk(FУ.?;oZzyvVRJ ~DE&S&ďd"w.aL&j8"J|E&Sp˰{ t(*@CӚ9\-.*-@~5nP>KRQ%fw>L_.O}ohF@g\g][[d}}B]@osj9:zX2Jw3Q`tY'7}2caodY@;- Rv90*т;3maB\-GXl fz~Zke{M)C0!B*CS띱Vc&`Cx-#chnDdU[G}>_חqjтKc"Z{ gފ V1Ug>jd2 XPD|0ar{+%R.՛"U_Ӿj7E['<G p2ʛa[<9W+UT. k ܽ~,YY*>85v30"T]՘6^p0+j- '|T\YM*Ffd j0DPR5c6r6k(e/0gIzv9w4&ꎳ]Ru2(cZad`(%#1 )DXҠ0I XrXM'iuKI;j ہPl"3c+U?9Ȥ` N"f VZDwD:{RR|ߜ alܦ8L0+&AY[D0/ F/p*um0ӕ_!|M6t΀$,Vƀyq:J#68RIOjRz\TknæATGoن\ngZ~YJj+nCo0J ߧеaomԿ'[6,rˁm~7~梬]{UA aÅ_vnj/CFhmR R1ɑqw*E{dg K/KǙൈ{!"8"<URu ʔLz]dJnVztYo*Τ2bXQMD4J Z(hRc|gv&jgsŴsD0KxJQc UZ@-Q:  l53h@S)iU-@܂u̸` ÌМAdF٦vi:l,_^9>Y>ɻv<E3ۀ jLD4-i%* [pOo |IN#-lJU}W'L*Rᣟh|89@? ] {bNjۤD^1w]xI2.ȖR*@Y <P00 .FDFq|~ u #h\"eq\PRȜPJie!I(RynaPN,͉94'ҜXKsbiN,͉X* gyY{gyT@*bgye_gyY{gyY{gyY~4\QFxYlg=,=sܳ<,=sܳ<,=OvR=m,l(LFw|T+ap_rթ"H˂j^犓lfd-WQ ,W#&0uB)&J7EO%"dUlf˵rMV͸!cW[eEbVbNkcP:يFXtFc'4+Ҡ<QF*286/QTGt2 JӘG4Bk3eF͝QFGR*:d:,+q3z-sOVm.ڦàvE,cYJ$|Jڀ`NCD.OQ|De]ly$PUxr4zU`顦Oeť .;X0taĂiePxl~Dd (9, ʺ\ghä:o0"m3I'n[*/9\7 o,g͑%{ofhI➲00.`ȯh"aN(E f-@>'ңUhĪYV͚'C!X[%6"bQJĐv(!' +CH 1ysmJqTQ(1,"1rW@D huFf`N!3!GZE yE5mJ'?=Z(Xfs$#B2 uL]PJm*6>ɾ4;˜Y%/2[O>4mӐ[.o8:mFZGBZ3dZ>PpeSHBjWp~9bB* "5P)C`8SgQt9N9 a hB cZ*3v]K Ed2wr?}=h{sQWFQ# \Bۈk:$rK9N'SC0u b0%($M2u80`INAu;"E Rr * J3K`@0>WC xqEI_^&/LWCL} "e0hy%s2yۑa@rO!wݠK)(ԝ vpwVN@ ^.sӟJ`V 05!O.׹\M*rLH]4؋Ga8Ic,f0||u]T%Wӿ]/_Oo$u?s\NHmJV\P&W_܌g>:9z=$F0!y|yU,3;(F|;^_lja!2$u$Wt6 i3Zgw > GLVpp7~upJQ6j۳J='>]m +Gި,iꦄ57o֧{szw7oH﾿珿w}} f`\Ji#an6Ko; <_ᮆCSŶz76[Q2C[}7[Y+}/~R-MJ4?)tfUO+WlYrU[JszݵT8!EZrV#2ocrޖ1[S(CT7MpyW =؟#2ޚ)jVɷҚwd5ޑw!y(` a) BDFq<'gՒ:ӽ4nPrR_JL_j}A'W7]`eMrnbTL]Ung1n/Cڢxv{OE>`WS9]&IZ~M]Y@ҟ2¥殶K۵ϕD XCe Eph8L, g\!Y0`O0,,-ru~dBq`X[ Dˋ S Kozu줥tQ_v v3pa ն d-DQ܇0Zܒ dk|e:O<|:un6^GG g^ɼס3%a3 n4 Oئ^z /W/W/~#oM6Tz4ۉ`Y+@? LZގnR~7Q~Fy4L@.*@Yg_Yf.3S!/k\|Lmg[%ͥ^OlZ!¹`ɫtCePWWese8`g{7iv<,k^ ƐprSYDػݎgތG_.?A$"Ga|fPEU)qY^t`$+Zzɂ1 /dau,HTNW$ J9gp+kO62W~99^ x$ )[X1c2b=6MVHKDgkgIﻷ?o#тJ# baVQ[A8*~GB(~XPD|0arroDueWlp9 %4ӺjwEZ[O%<6O$klb ipEKPB y(m.6ǝByDŽĜ`=anˊFӐ? JڤB:yI=t$LEt!"'U٬hSڝqIh%I2"puPXstcD\L7}0r!c2 CO88$$HeH[/ܑ{ҘaglևQ?-HT4bgXH:ֈ$kĬPVZad`(Tkm#G/;L|? l nfn0~ >mmɫx_,lYRAXY$~UE4(%LR#ֱ6V6ێzN^ҎZv KI Zbq[uZnq+uq{՘IB(5,Ľ&r>D T*o&8`a"XgZ\& WQ+J\5upcBGW@0h*;s߁P19tJTR +yAU"Xˣ+ W⣁D-ӇWJI:zpũ9{3ҼSTL{/7CoG_ П@ x=Ft$^e"k_nO!&pOiwUpfȅsc̱l2ĭ4{0ɾzMy3-PqڨPSW%nBz5|P-V &RǏo_ ˵%d81Df5<2mśgM|4uǫEǸ.'q 2JUǧ隘2eYj:Rj3fȴ1K=M0&YNd7m9eKHY̘C:,vAm5Bnw+Mt;kaVE>Ӓ)E 0÷>By*M0a$MڟFoOz%Nu0Zո+*U}J ;GQ|MF)9D.;Dd4Tni, 0%GW@z׶*+UvSzha-\= \qʑ>&GW@.X*Q(W2xtb9ubwbVN<RHT^2l4歙ZGJ-S2b@Yc3Jx^ EB3惧`j,sDܴ@0@"cD;։JHZׇP-`N?f%oocS֧;2% 8!_(|;Ƌ9cwSv6QD:]ž@=hNYG2roTFQs!Tti/ZL@)n5 W)l0Xc*UFYJy>%&zqyrzY1*{|*# QwRsؐr~ts|kSFx#߆iyP;rBI`ʵRS3iwAyX0LI@$% |́I:<&OYqޗfTxGDJNHd 3âҌ;$$" (O X~A< {vӾd< :a} +eo).HA(Q#"KlBʰ0S`OϥV$1.e GJf3EJθgAHG0E).M +Hkm#fX٩?k.Mre` ={rH&\Iot[V]X>BIkEDY埓Nj.b.q<0z<ޜ;2i)b@Jg@9FRP;5y>i>|:+Iy#xN؀K i>tir)* r H])}°X|O#_|'9zצj`d_՟ ,:??|>}19M[4Xcr$vl}*_ܛR6{Уt,bUͬqk˫rdٙ"`7 ȇop:0|jhIO gZ563*@ } L4h8:o?i?պM6U 9)OYB-!aѸAb/eP:ç7*V͡XQ)):5T*'(+=׿v?}xw}D]뛋_ $8P M$̌$eĆiCwSZ{?|}=oK9wLE^gWOFW+@6ro 6Og_@ b33i]8έA6܅>C[bt6[2Q=&03(EwY2e!EVD`Bd8Nۄje*CɫOlW?m1}}]̹VTJ]Ȟ7&tJ ]wx+%AF_ٌ)/HYt V;@H`EȘ<3,A7wAEm{kx^h rqɕ <w577$׋t'8BXg"Ԧ0+fRD[}nB}}p'KU zR[r~IjiO=_rw.9DNn:9HE*L<)!SDc)(DpC@J,s$#mMR |fgx~_rjAPӵ%W"xU ,KC_R˼Ñ$  gIh\Y:xfXxDSKP;I.#j-ڶܷ i %m% =^d}  ;{Jn,OWb{YHJT^k]}|4/j$Tcv{E:)v(6̗zZ3ufUp$]?Cd*\= Մ1E%:B=݂ju\5XT6k*(DDIg}mj}U#Wؿ<[xB7eEݎF_\vH J`P\høYb9ȵ&hǰY+G, xLj .7H&sn[׷љͦǪNRCB^w 1ZtrSLZ#ΌA8c&L#&2TTa|t&LL,Tkݟչʝ+Yu#~W4 B7=_Xy8pu!D!e!X1c豉hj4BZ"Z;pz7}#]=H\bP{ g  V1%e>jdU70$5I b9N`.0 r# T)p7-VJ\>nk,pnRƠB jwYҭ_l_df8rݕf-!OUՓU.*j*"?3$>85v30"ᔜИ6^p0+j@C `D}@$ mm渓B(' ~0$xTH'pb^ew) Ӂi]I~M[,Gl;PP58Ԟסvg6p<lq֜(l&*&D:*pS.\7}xY1,C0pp4H$HeHw {Ҙ5ra [MCxTZt,U-t¨P0N^iGjXRZ{1[,}ZUvm#i=RŎi|l~ntzQ|B) Kj߇b{maMS_OLCg(M{ڱ3opBroҦ 5c`C=F˳ΤړIzjRU.`^*IU74K"BNN퓲ȫ_{0}xs<mE h2k ȥ9v;jŎ6R{7 L@. @YjT,PșZS'1uCe$IGE,`2SZ}h5-U rCݽQ`Cv:R>*qͯZ7POTWx{ cEu3.N{PvJZiw ?%z<g,Wm,V_Rxhs2koâ,΋GT>z/%Yz:`Aq.ȄIG%b˾Ps"[<&#u(2J YjU$Ĉmo>H;uyښQ)tAM3z-:Yu/?Nߵ|܎ϧPXr`2*wY)ʑR1u)j'JQsb1Q[0GoL&.L6A{,+ - -)U #BR)*+VAv *;L!zV`KeN ]:8/jz.gv-sNNpس (I $J}?BxGV*x; 1$l[ `Lm$&M)W.E!T䁤h3%$IR@wIl$5JyU?VZ!f{_,*]Qm#;|zǁz^I*y06 &z E'htbg4&DN6M:P*/cm{p%gXCS%A`'=`+V"Vp32e36W>Kc>E>)$+r2ueWtWZdd`תTS>˒@[trxԚA-8|fRl}&J*/sZcm\*Q=e0q6[8 qkEPV=7Z)Z; YYj(eoK_g__&?>xe.M2faԮ_uߝW`~Dι^ST=VkRQ9cte#T)PVVR\u A*HO7mEӳ%D/ ,8n=YάfyYլYX+ 19Ŷ(Ξ' )UI2:N.dK 3%a؀AThs1*Ve  Q˨"},UDWig4eh`48S/t2* !\څi̛-z;|t &G AK7}}Y,*0Rjj,&-SԤDg)`l[̲B'T/~lfۉᑈS^ğ Їv( ij2r.$= & ɱtϙ[Z!\TX,Տb^(RHdL\Q$%0Ϻtpѫ{ &Ζzŷ0v}GM-*a=777ڊ>KoqÇ(~揳slq%[wo}}s;g}j+\+:~ȷM:q u?Q0.OjQ]-^8npˋinXJ1otNj7| Wׇ6x]?qU٬K9KLO'^n)j/35Kx!Q8/oö0XdnWk_Z應 O/%_eဒkQL~u 1-K 0*0Ԧ;{?QΗ@:gpX8:[7$Ϡ (-6v+g 8?({[O#3D'16pdzok n\w20I|5BlǙJqNddEF!Kp≠uyڦQ)3f̋Sӱ9aFxvf.sCݎtz$|g&r; >@Qc%rTLY#d(GJ)j(E9Ɗ(Fm!h,{m23}>@8ʪ3RaMJ=ϨHQ\Q ڴP1` [B-sg`t;8[4Io~ Y>]|{AKj眓UL d0@@vG.WW jσx^9}$'u-j^a]DH C*]4u3\a*0U!&Utc(n::mV1U!6ARX@-2]6Z9)Z\dSO>zL3\dĮr..nmGr  A4GYMIIJ)bbw>'ʕ\M 8ͷo^iq+k;N,J7w?k>Ví.Ug/7S%XiBݝttAr{*0w/͛D@g=~a (I $F.~6 W*xK]9[pH>G11:k$&Mm/E!T䁤h33{ָ7P:mt$F)g}{J+lO-bPHPZcO^WkO߼ilD[D>Nl&Dzɦ"^Je2z⌳k(}jC1Hl {%Jdժn&_BSlrIrL?ߞGȧ>dr~E_WƠLZ*QLLZjgYHvUcNNSqZש<GRXLJY"V τ=S Y%rNkp5%lقYo[+WOa X̒и0o[7M\7hMa^0{;/_m/Wx>D , *BJNQ]>UWl֊BLN-xXC(V% G8yֺm%N0,?4J qN WQe  Q˨"},UDWigu`~88S/t2԰*ڍ\څi̛-Zi6|t-&G ׍ǂKtM3#152θ#o,ߊo6&XpT0Ƅvgq*u ᒢpeEzy >ѥ'{mB')~lfۑqg$+>@l\2 ij!&ǘh)@΋ m;:::j}Ծf7Z?/_C2 ,LUqIZH%KD֠)|oq|>Djٔ~{`վ\[kWW bui=[=5Ș¯ælj|,+W2g!֤BQiIBNRY$ٻ6$W:b+M}0B!˚cmτe]׀|;E.8,`"I JHQD i#7wnI*EM] eKuiWݹ}7N?au te{FDe].3naHdDU7[Yo_ȞL5'npo`+¾GoBJ}"pFp!VisS\x<ˁ?Pu>7̇&1eQY.|6fR3q狛~hy-7G]5 vY[vՠ }EnQ,`Y/7-Ƴ.^@~s\C')aE)yj$ZfpME4:8\^2WHf/`>`.yֽ6# V/Y(nCL廊j.NRJvm6ftHvqI7i*?QPn)#֕cCo쯀f: w5I;5-mX^CQN򻇋׷ڼFZ3Z =FFLLPf_H1xǷުŔ2-DM&z5hbkTZP։.ZL;+N+65JʑYQ~aQःWE J`P\h>շ= ך d]F8읤't0z`R{#쭺<Ɲ*Ѫ)7K*.l3 ͦǪ)}#sPIhzZ,x?]%{scΙ2׈ %n_}@j)69"EsØI""FeDbYRbosN@m_,Bf&lMz`/M`T}71Ĕ,#*љfS3p遗LbWrK MB4LW ਂ1HI}$#>˨c N{Fx>v1{#{ĸ;sxxe&#QHYubj3jXDe4zl5ͭv"G\;Gڽ]L>=H܀8HfμZ 31"G$252 񎫀0$5AZX,@(vGE`.0 r# T)J HWlp>)? FfnA9]d]:} tj85v30"aƘ6f (,e-"P2zEPGPzSNq rf7 ,3Fٌ뭔G8c', u{OHϮSd!ͣ|X$iNoPzr3?R6EK!<*0o,0wRs!ݥS42OP` NK!#a:0- 9ޥ늜͈m#\ jwz-Mb;š DŤHGN `ʥ c)F.qVqC'  D9R _ xҘ3r6aY ;]tGxT%Kw錜ɧs>?{Ӂ# Q2p9ąfǫdP7mm|tNò-]{ wdur#y"8E< cAZ]P , S2xI$s`@ _nx=MIK /6C HitL*"SU03,*͸ `  B' Ll>5{m+C㞇Q(7>]am2R]^IPS%F,D%N0XUNQz3r=.FpZ{b|ƸH@08gϸgAS`R%M u+Huwv 01?ss_/gz0eHN10hUaKcƽO/%q^(4WKvE@_2yGɣ͖P; ܁?EZ\s)(ԝ4\t0xy:*I@"xN؀K i=ts5Q1' ܻbat.Qq 3"|>dWj`f_ԟ ,:;;]^.~39s'Kaʑ*Cs&k)eh֔*ɷٝUog)p//g&-^OQ/ʙvfEh_]|ja!Г$'Wt i3Yd7 >F YL|<=sp95ß^ Zk=Ȧ^*f4M`1xRS@-S|qObE5 qW_w_}ï߿{;LԻW߾zuᤉܕ_D ^{S)]u ]SZtor/> ?~ v暵 Jf?]z9 .d 5dՙѕz3_1,ӬTiNo]BXMqijC?NZ-]}@Ų͖|cPZxmX?bk:2K~V;GJQ&3q<({O۩>qFNo=lЁ; 7[DkОDe.=ꝺ9/DC zt$7 sWriO BHkGzBpj-X &A]xO^dv^|Nuw>7f. :eGIԌ }{$8;\\^G yVjp^M JQSAU@)i>\p464.MO[uڽvk74}y4fŞG'sqgئUɚi<2Cm)E){ΠE5e*ٟ%g|V$6(k"x?*{S/g ѨdzC憩Y'aR$e}.BK{ éCf)bqmr- ==@B3B} Z(-@}SY#ɼՁ ,5q +֦=>(~^h,;]e:\Ǣ. GwPq,*-8Ȇ5^tJ? E<Ɉ \%bR"&Zu)bda͕iD7Z"Jī$ 0I1޶agtu1V6|Ae_2d"x!|$ rTL tWLTWL}9WLK/NZ+\q4VMrm 793Dn59<2K /&6DžD.VF'­y~%^/C {[$ꊲw *_[e>\ VC-ܰ=6?}]Ȋ{f!D1ݘw*-ۀ+4,w.C\GԶ#XGԶ#WǵԶV 8vT.&DoXj.,;J䒣D-WJAzzpE8Q w_\RS,{65#oǟugFeX١L(.M7^0;n\y\{@R/$,J~K|MSuu- JvPPRShOU_凬r1/g+OC1ҊM"hϛ] oαȋE"ZWXR;GbqGb'r$]Dćb'*u|.6U ~Dpfh W \%j9tJT.jWOH#+ #u4p%X*QK~5QTWOZu5b \%r8J*}pTr$zzp%XiYs3rNo*l*;0SŒ:k3Ugνv@sJ>K,= O[1G$eɨh:sϭ+ٞr.0 ِL5ƀ`3L9P{qϮo{w~ؒG?}]brwڞ^ݹW4ҦH9l ˓,$Z'uЖ*;q?-^_׹c\.)0/0֒Bߔ)ߞ_=3s!2795(++?å8$hGQntHZ)V}-`Q8?Q>vJߔxm*ak'A1DV^W$Aw! =1dʣ%Р(Q=8cO(@$/5ON`eQyRPyghɼY'yt.=;VR]ob$ج'eL?j/V_?Ǟ8_bV#z.vtҧ15hciӡ?ut)L|a';-a_G˪4It;.=w_c=,\J n`Z A.-L%ʬbN0p5i+98 Ǒ)T63I—6\})br N0-Sd2bV:.)*z%x=rq>s d6ٶU5bL>/igœĄa.,.]z)kZ}__W(S pJQ4VΠAEc4t J8N4w$EWr -(bLҡ R-W!F!HcQfZ\4r!T{*Yr w g;_AL>g~ZO>BqPtw8n|Ҍfe&^ׄqLGDB 0>$d:# rNy(}&{2ӧilbxfhX쒁v͉n$`f+7yq%CyUZ:1;*+D\zlgmՠj;zfÓjc5u*Qc}R4RH.('1 TT. F\J8.p $ϻ6 HJ Sk ?JJbNVQl #ݏ!>j̛-\)D78Wnb~OD*Hb╢>qX_6/d4If&mHurrL8:s}\_v@J)'RZ8u^'*",W NS HZE"1( w, D3qeV!na[ogG~</Ib==%<̥#ٱwˁs(%oN:;ֹ-? 뛜G1 pEuiW^3i7Y97}99]pJ&fZ4(18-(C Z(OJ_J? z,Anz/ƿo{9$<æ4WLERDUTp2k$,,qpіh.mL(RArH}Lt$OaIC 6CJ/: :Ǥ@2L\)mn*H єjpchg1c#P(MRSSq< :e]R֝Q=ge9FʍJH$iaHWyj7NuZvQ$=ss!\JR FUx2 0X`pIN*>&rlϿ\7G{?g W粧/D>motTZc@,1M`ޅ S=ev>% oãuD[1'e5hda'\|ЊsÅka^}7Q -'Ф O׹~p5n| RSHu>;Ğ ]~ލkd:K8og-뛥jVPT"vP_]ng?}0~%1l õ*Gd?RFnDӻ{۳mu?M|\jw`PQfoa?\MfǙ]p mr`Vғf'!t kF kZYQ # G F Y+h8xMtgp^gk۳jEţ$yf;(>CnFJ7lCX(cbC e^\8ᗿw~]7_w?̼w\u3p !׭ol; I}]sBKuͺFAomwWSA0_Yw}3#f Otz\3*gҕz{q6q͎7ꨬ䏈c!~4/]ލ{Ι FZgrLQJ=Ai`8yѷ{:6gȟ`@Hb VFb A[U~\S8S:S>}y"T }qf !,Eae#KH/#錺jTs)z|e3JjM4Kio\ #6dC0Ark6lBp{ws.ћdbͬ;`c{{w+꼓=^[ Jɀ)ʉ@`eQBϫW/3Mz~%O6$ɡhSD;L,pFK>ieFHe"µeYѪSx8Ϯ,TTɂ;g1%&+fmv|ڭO?utr")G"D6' ŠE+9#dCBr9ݗ&C:D8 "dY^3C / D F(RH@L@:G2O82: ×[ 4C:C@~%^I}-txfE[FWi/S";.1 ʉ͙rA! FDM,W`ыhI&7bxJ4&)OHL)]Yo9+'ۼ006{<.AFRv/o.dRfV`;(&CLHGH>MBIqb89M66yP|غ91JI&C 61Tb'X#FdY}!Ύ[Y*|*ol3I϶mչfyP^SY^_g,Seyh2kgβjW.>17@`ӗ^?a>G/<W|鷛 ۃV_Q=ߏu kFЇ"uO2b:2'4~lz6]WLi<~3}<L}@ #`j K:t5iœlE[" ^anNjgi 4.f7q~/XًYwx4 ./fYYԋi8y,_nˡގo>_^s8/oW7Q5OJ}8P.A$/0@tn0XK4C!9iij84}fPI} E56KBFfXD |.1-e&H"1DצlQA&#u )Ei IŔ/RPD &rxp5osFF\}{v޳22Ux[k8ȍC PyvaR)Mdz]pgzҵ՞ֹ}zM՝/n+7fgd~g[G=#0=x<ϳ6 ONoh9lJ ϚӚa+Jⲥ@=Aa5Π.KiT'TVjmtLni- vy5%OSBٳH2,l J٬1TkpV-[m ;cLe DI:E4HӥHk`D ^88kĽĵ>Dethw1C__3z~X9.Ho>mOQ+|z'zNu,bȗe ώs1)ct(}x5 3}o&0!(TrRg%3Abs!r"=N)EV5FycYKlfh50w昻DKw/G{^{!*1Q.j6i JD"(hxآЋcu-ؗ%($6( bI$S YB Wfd5pk8qM4.kr]ϽYr\_31K(Y>oE 0+ZZ36.xC ´2zZy0Ţ`}w8UGdτt;VR^<л`BThƋ$u̞eB)A $!wx8"^dp) TkGDJI ]))eyCUl2[N#`AbW& xV֛>֫+/޵uwu"Gn]#zV[=rGnG}Gnȭ#zVwȭ#zV[=r3eG+ڮ*8fJVjsP(PFQC4큦=д@hM{iWhM{iGޞ2'ZP.AkjؖY{ .]|iꅘ{akĮe~ݘOێN}0k??0ؐL$Ɠ-dE ij.jJIN䘢BHk),}S`ΉC+g3nsW+|ĞS\SS>f,vo<@>.if+2Zq?m];:::2Zx~O/1Bi?pH$0hpCma@Uáe/~^Z8VU_ΪCy9`;Msf68Ĺ9*Ĝf7me{d|؇I1-cHI`Ti,Q Q-$/,!` 9~ْqq>s :|pM)]1/NaNDaAR>xt<ԀPaF2 A0=$`pR(ϿQ=wOFȖO)K-QZjj1ؒ1beWzSˈS(~ *ʈkek,(ƻI$!Z0jm6xm3 >Mӛϣ)dYſOԯ㥩Wka}MVųHBʅD4" a"X{W&;GodOf4SNezɖ*kfŮm7'7Nj%ru{Wg25 s= X z &G7`@!AR(q"/=Y : Y# Ov \dLW$P"#bV)'QRzJ*䢴QԶ~V*-dPkS-HWLqجDIB..E[r"R'̐RƐT1"uA %Amr>=~E?Y)&O郹vkYtO$Q0)bt\M U ( 8.IJQb&[߭!2I7^iC`b,&↑LVٮi<+3eF0Ѭb[39٧b3mZgJQ}6LM+)'r~n˝*},D{MʦKS= )C2{i[+ @H^ {"gg\jOdKS/%%Q0GdC9`=YRIAJkjm ge܌RNb I^pn j #vw_kbg'1^7veU^\_<\_M޹V@jY\b=!<6TJHoȅA[6,lV`IPH4TNIiUͦv>;N =bQ&d*"zgp[t3'\v38 O I1&Il#}DN$Y&qƲJ0?c&%6:9+ 3dF $eMm N ٱVx,(2jк_lSR \,b3"Qu-F `Yyd,&NfTdK2R\41NljcB8o5ZˆC7T;0`ȑ^Sș2^άlEl6R._ub8uNfR]4vō h)ZQu $*qbDAQIna6}ml~&lB{ML&ϻcz8!7\U]٨+JulY`$$GK1 *^D8hS||]p3B^΅THaJ)ǔbmaQj^TJ )O),;@HZ[rHFϫ}Be(5o6#1r~#/ܑq^aqz;]mWA8w dqkYevϋ򖻐=  TڠrV K3C>e$a EBOm:ǐ!_Oj3H _rڢf).AL ҳF"Ӷj3i{g1oǩ@~QiuWe׎N&"r&{$( Q9(%IA(wi?Ҵs{H רjKg{8W!&!d& bB?%)R!)+Uσ")!EJcivWTWQӅL r}4)mv7҃Lm*rcY kϙɎ {A-LeK_´7ʥJǐ` (Kzߤ ]0p}0޿?n]')%yɣD|-6ý'Au 8aGT+)Gn2s?>LQ _PsU*:|w W3r5.H?ԥN2pɧT}Ώ#s!\Opݻp(W WM|ׯ տ DЩt09 4oOqI泉K<+βJ|#Œֽy63⯥ɻwӳ61J(uó0 gr{s"dOӲ37-o!+FҌIȻG:_5 [5֬fU>602 W0b>GuГ1IrTF֏Lrը74ɷ PfNɇw'\)\y sP~}>[CO mkhЏz|q%is+ƽ}p-~c?'_:ݜ3z{ ԖO(2UVc)B "UvmD:"] e̖T1J'H~;eY67 }~5?Yu1M!~6N#X1\mٷ+c79q4 R?)lǔc_?`ɶN#ԛ> *xp}yr6<A()SrRg!@JǜW+ǩW}xҩ?G;@)ޡ릋~CV>{ȳes CXʆZuoʣWu~K мi{L/I%|_?U. ^ymըQw،\%߯ޢBк{󰘛NSnʣ7:lz(pjQ(& \;)s4UDp>$ T'3IZuV$B'Kw-Uת) dݱVo.[l\n0oa)f;EX,l$#sc>lo \v7*n:[q7bV=;ga{qML;^3P8.wsp$CerQO-6|D僠DO^jbĔV. *Y/M03v88ٳ{M5[2pA~rpS4@|d<3k?ny~7ŽaQDZ:IX%2tZU@gkBfO E ^Vky~z; ?]~~ʘUx8罥Be1ՉKYJ{ZFaB]g;wEU!k}u} ou{En\KZu睩+.\BJ>CoLcϭu7=]>\[w |V7۸lzkƣњ |Ϗit7qtp^0 [.Mp*l҂*娢Rъ$)ةک{X̆L&jL iTg`R2&lD 8#jQ6 qڱѣ*pJF iݶF&hJ TYsY*1xEKg]h'~<")UC9oy)N0iRqu4)@EQLEEH}Bp (%2|w<᛬W';_*;Lĝdmoa"N4D7J"|4X єsl ڨ<QJOu+VU_\2SogpS w9@3.W(#1"GqK9]>Wci8:ɴUw [|д|Py!W-#\6c9VcNzHEPe3$X1y02\JeL^,S)MWV6cH#H\!JW\E\!}WJ։(8\B#2Tr*S)I'^kH\!ڃW\~0e3>/2W/Q\I* %$2Jr(*SmM*Si'^RLJH\e>q*r0`=T]Hq87tTYqe>޺prPdvߏ4r8 )NG!:N.xo |$͙1,6HiPws"g޲A$2c^F-iyw߂o$)8P_L!s E ?\.PQusyţg~4DD`1@+8ܑ$ 'FPWd[PzC]o+(9FYk/ dZw 4>yE.P_Wge>"{&&^!!&t yiM̰.P[=,lL+.(E :.;m4(%'iT.0.`(Ц5D^G|gIeYGk5$l'Z@;j̬}9FQSu`1Al`'U4\PNjc,AeRRƹt[$)N*@=ƘDV60jѲڧ`aiNQA%݁,݇`}0C= H$7j&|V}TJ]@ H = 0-66sQaCͺls)&NF}p3>?.Dٮ僘?l;CiU۷/PH tFB-!b,c !,<W H8muN?OvB7DFTuyB<."H^j"zˢ9d:l <19@]6Z>Jŕ^܎ %) CvxuV~ء'BGZ +U}y7 5+_z&] *Z&DKF:8d[K=lmRovUtFՠk7U5/GpEa"˫v]h\(GuhR<J!#njR𠢷-I_ݢm< cBA`עry3A8p(sB蔤jfk߽{\ mg#P2w6P!j{ +5{k NJf, Tʤ@.$8Jg`(LhU=Q84LjMeDNpolg<T&cޣvCS$m4}GZ3%}T GWBj?C,d6 hy0:=^Zc/uI<KT\D$''hbS%(g % Q?{Frl /Ib\/,K cd8eT EZm[g]uTw=]:;_Mu5@]/+q ֒ a١7T82؜ZFCEe- sPTu]zKY,*\:2!'LH$iC.Wyj7NY}=liE3!ו@\JR F 09cS`%bҤcHeic듼[޳Lg|}քGL 7{.Mvbm2mCr8f sy/`r~@ ~]0‡|SVzeޅl鲸>%؈xsDyQ!4a'dJNk95\I=G8kFpVBˉ:4;S@u/.'\݈q@&9M7JջOp=JQ‘}m;eaJ(*tjv:'aEwpI76_ZqKΖTMf n=Qw=o3vݯ7ͅ.W]aPj9g/ܧ8ۑv9+E8?]Λn|3#lI3'!bS7lS7RfuE504#G1z1p$+#{]<^7=VT,gR`!cϓYM_ֈ^iwO L5keLliT{p^~۳^ooξe^}:CDXCEAhtڻkNhE܈=~Л4{+IG]A1hfemw.?}7됳']\LsLskf[@qA\|e&@m, tM4Ry]*+El 8l!mϹUc\e\9[&S$ 4f0\<.McO3I0 x$1+#1ˠ-s?K6|LPy"T }qd !x1e#KH/#鍺YsըiDvhi_xhѺl&QIxs z D[DUw#ĭ߶H֝m;-{3e{9nzE"i[9_h^ƾnhrWWk=ွճ[JИ(o58Vh"jR2 BJ+r"*A$tA#0bأP7^d?]o/6Ǡ|(IIZ:E3:gd )PfP!"4P!?Z3!.͢tLFasߞ/oV0 ll]K5[9%祍Zxf/$«Uԟb?T:x.M41ыhI&6Hbx3r|ZKVP@)k8d1Dz19JVV8Bk}1?hcRFYoS :/(A2%gҖoBK%@<+pw&BUO0dQTA*F9 p8c$!)2"Avz/kgxUԆDJR[R8tzyH=xꎠ`@:m[0)m޾t1Ngݍ,R+?/7 Pw}SkGcxOOy_NS7Md0r[:ZæϞ;e B_=<_VoK͝n_xM:?*=>,壚K:/g@'"݂͑dmLSF$i^P/nq.]A$IJ $J")!d`$PO1&$ƿL+Mx*n"BW+͵-[u>5!W=9L6qZZb>KP,Ye@+,7]o^4zV`ɷz3~o/? .& xtb'i}9/?~0iz\iQ4`zHGr/Aѳfͼ%  zu/ؑ/&j<}zmݟifXk7TmK^ -ۯA/ŋZ\Ll\;A{v4هk7§'Flqv^^B>VQGb;9ܨ9Tnǧfu}>-`$F;Jm*2VD Ay8q Z|,93F-(+`A$$4Ed L hScp[*P y:Q"u;˝^iTa#8{oc:W'+k^٧o61gt>g''b*+k š=Z ڲ\o*ԀdȄ|%I"8hQI4#2z 8g3 '#9gM [׷1:{rڴԜ)׋C/%%IQT:M$k)I0Ghu%3Q97Iy w{"2:g$p0ȟ(CxԨgEѤ$BEp A42g72*Ű8 orBcǖD/33^lw0՛rj4_9b3!5I*$5j#D i 2hL Bi*&"%Nnx8 s=,&W:($N"jfNhcU)@"A'ٍ~2+&池vq}x=CNS `̐w4ɠ'$RoRhQZ6$I]FR22C$OXŀP+!$Q ڹxXLx ǂǡ #"Gĭ@%X1 GV$VRRVȑUhQD@&dSL݈xu=b>bP\q\Ȅ\[AX o' 5D@9K yS6xx,xXlv0 lG6IF "`֍q w~&gma,x)qޏq^$jwyTB !(pQK`_%j)` p}ݷ! m(}vѯzևTe$ 5d6 -uXam;_7{˦ #֞DeRGC>ΤA*9a>@h"")er}eFu ѱX7Toai-LA&TUyz>n&jHrJ9TF+^xRbo&jo r1S31ܹ`T&hg`$Rh.1Q`$]He@M=@ܣu,d! d"NXRf.b+sv9g ζO^QOd&0j ghU҅s\:7}dvW5LAP)CSBJKӁm5Tޡ=#axʲ(%;lNmn"ZeCN㞲Qʄr1i QDJ#Nf\̹H@rȾ::<6FVQbAM%gqɋ,wem$I~<"i1],0GmT~#HYi2)4`fQUED9RTYe)䐠@2fU 𬸈>s ˊs Ϯ7sX #im: rGO*0p>{|4Yr庫􇾪TF(s;;uAe VBx7*pW!l(g1JWe>0I&8$$Ι+ɂc\+52M5lĶyMz&)]_f?/KAr[;~.cmu+| NI. ̅ n/ nĹ)h">H'>A!{H~3؁^FQ;ɒ LYd946dX"e]U-8:r$FǹvsEe7m6.)GV8-8q7}= O;"K{ >y"'NJ6cjsD؏Iiy%$lK깴 UX"eCqWEZ]^9""qCqWEZn])i5+% Xxw=*]_L8ɗ%~\8]҃!^ O~qz+u_ofW!zڣgy3>$ޕ;.;O MctO?on_0IP[GWNpi/]{?S3OcBB* l!@`Hч8rYJd.s>'DMr8;|K~ .\ma+tx2] }|Ri_pӗ.WN/I sJO_xs.ZO љ\[wS+ߟsÕW|Z ->]M]tJfo|*}ߑ WnQ/NN41{.v}P7jkZPuqlщ# S=wyB8ĂrA(IRPve TpHMGmMycuZ nKU[g _B̹.=o KL>}$hMFL@rBkG.V:a }&@Fަz':Sw+Sm0'0`GK𮽔~}gh0SC:˴Ӝ CCn=q~7S< ϤKfgeBJAGɬ.e *+O:\7z '16ڂk̅y|!o7Bɺf@ke6!zfaІIKK`)0̃ŠƾXx WOy>W h֢5H|͉LM1/+3)E\m2żHkٹO1tm+bNO)~~뛻M;?"7}$L2Ct< ]At,O); CJ"% EϐyΎ0 BL{GusQ5O+Bd̨WWrI/T>J1T(=.PQwژIDCц\Ug;{Б2[zƻ)isڛ~?mÉ_R;R3ZZ} OrWx}IN-2뤔 ӝ)w%sE ]Wt؊X=ۈ]GnPFv`$mNi9Ty>Vy?ZfI87,Kn٢d tR.J#GPZGڼ@sͺ,GjV%Xsn[at䐲CFdVV\`:1TvNKѫfD&u[*oHf`NwzyDd$FHEp*cuZl7TCsSMO *9xYxcq>Azݠ wGA1.R,`gm}QNI -U=&C`L[Kz0h!; "yL6g,grR=c5q{zJ5_XM3e=_/[00-8PdʕYt(yZ^ǀg")*R%c9kb$5G^#:]C&"?RPF#=XH+{j=ҫܸ)YMKP/_|"4Nh!M'qD.E7C'l0Xcsմc_*C?< [n=|Zӯ;Я[(P~Mp4ݏjYqaŌpNg%U*NEpzb^iDranā!x]}k5SO7xu(f LXIimsd%-3"B`8 U)C)RY })=8HzT !sr,cR_u"uhi'?|JGb}zP&Wo7i]^r!eʥL*$L!Є"Yk LdTlB]88-bfSF0FYsXusC T'oYp= E{c 7NFfkƺS&>iS‘.G4Pȥ1̲&ZAAp-HԾIu?ҴS궼{Jcf@p]A% @[E%+VBjTאJzk VlY?6mϿ̹_"|YE;Ṓ,9+|a[$_r<ς('s[L/=H}d㌳ē ~x'KlZ%Jz d]^-%O!^N ~đZЗ0x,Su>.dR/M_\wswEp|W\|:MHwmI_:PCpNll5,QmLZi[}gFр-uUw= WKuVŹ%oӳ+%zBQJf< 4|1~%1/^l^qk&'T7B(QӺcF:~zlz|rY9IJ]:i=^A2ٴJԍS;ѱ e$hHB>Y0m)ifUޣF+̳whx1r`rݣ_mԶwՊKN*%$,}~OBzV"z"!-%5:eLѩBG`T?./Wo^pϷ9˿+RQcʑd'B#,X!ʆ43{ld u${[Wm e3xI$*5/q1eW!p`IH J/:*a:;ϳxbϳ,;Cϳ6JqUv;^]ހ:^]{uz+N>󨷤fUkd|+%"x=a9h/A$tA=$sIw.?Eks}R=4$9 uhgt%s'SB,B2FhT9 ڣZ$ qjP-s#rT*g8PQ3#ƹDlrgiMC2zQmJ۔Flcc8wY}Tri^o6ݕ2~ 8j)"Jo#xX Y4)#:BVtHH.iӔ:="< "d^nx {`x`&:mL4ZF)BbJLٛyE@@]S-`MS`xG>?A٫GANK4|OGKz^U؁-B>5|; <,_ L$('6'D".M41ыhI&6=t<4z󧳅I,CA%"@t:E022P LF ?D]`#O'ȓu?^(y!h 8O;/d XK63.ƙ($95Qj1?tw& MB_ UբEf?P2{)E][v@lDv/uYF7r>ONJMIhb.&eh@\j `\3Z~u$dw=d*a ~\'?R0k"Z6 !RC~:>?#eG}A?&%% wEEt׼j>c?Lb.E.i?;Z- |ݬ@'Z^wF(wo+.M3SӬ`MuɦuWvψP .zKݵnT:fZ|)ٰwN>ӳufm[nޔct[r9dNI_*?q[G涥sFZtm= }; 7FhIjrQ[&iWe]tmz]˚ML<ۛCwq@s޼ÒglzrKbWVAĠSJ0Z$2krƉ&JOE{5%:9(̍h:K>v")b >2L5HI7<A!FcYH4b́!}mU@ngip%e vn5C) '݇?Qw_!~ 9K+~=c k"I#'\=K1bIg>h d+xwZ.qL{}НoH4'4eT@ҚF/P J2Υ :$Iu&AUI !#zj1!15eZiSmgn{Wz|fs#~k/1^wXD4[D*9j*fsEz'އ@=7IZPL7>R/Tr|ƕg "g u]MA ʿ{a Γ<6`rQQG*AjϼBB?wQzTK[>?|v6M*|m|z܁-T |YVQw~d Vk.d/&FއԻpi7o^‹WnC!p*NTWUN고<4C_{Q^7VRS@;`ÔWّ6C(? R78JBߚ>vQTQ~J|n6z|&rP1i1ʝ͆_XZnX嚩7)b猣ɀs= Mσ=u}sCW;mېcqDh5 _T\+ 0<3iL }D[* HV'S.yg+8v3r,7w'doSynྈ%1gt@i$ $ZA'&IEՙHݩ 2b % 24#hQH4#hV}Ą3rV jJ7Ϧ0ɉmIlnI:4,.ǃ6(iF./ݩH'Q[m.WfzR͔snIvWg4h*$\7˥%CppgSdTFx#n'z `F9,:&%4DE*XT FY*aag.*uvrz,x:˚qK>avu a>u_r'Gl0TMJ1I @E͡J[A+ebJS6)'Ӑ=32lrBIDl m;blb‚*H$hu–Y~O^jrM1H25SmθcS<VT+U{ᄭEu '}׻r{p#ν碑&h:=F*c?'(ΘNIZK ;=>K [)})b5[mY{jKS$vwFΚޚ >K|ݨRLDdzV g-ǫ!v7yQs]˲\p]tFfƄ,$)ɉ ڇDGT 4!jr̡J/:~ Y *6xxW10i-y BEI*H єjp(vZ#[A}b(d*]^ +܋YTEϙE^:2r%I`*+<5'@r$n3;5DP!`Teu!@cdD;M>[A耵5rʤ\۽^v1 ?g.MvhAX0>=G/U H0ĒEjNronš|V4a'~Ԋsá+K8Ër0xvCkE刉uhRwS@p9o' A8pGO2hZy WKuVŹ%oMgz?][oǒ+s$ vpCĘ"R[=!)j(Q LN[/xgw<9K镽zM6UHE59l`2t#~p6,KR J[CT$3׻[LOւ(}|}3.E/־N3^R_a_8Sq*Ũ_t&݅/D b75^0O~ [+\k#-+줙SDߘՂMSĔ(`w1,YT`#)E~ mYS&aS[+t1b%IV I[FhңΨ{ۑZ>qH7$uˁ#|I$̱V^)aO[e|!;` Lh q ǩւuNVPp5<ȳ.gwhg3z;A:1ji)k[=Mgrs[V۠\>4jd227 O-.."{nk>p XZQ2Hb`F1FiD1nY>(v{Mm&d(:ӊIq]LO-vm] 'Ϟ-!Q3I`!ZK23N|euh0Rhfc[ˆ;$/uQhm6 mԽ-<@hlI_jC'%} W/\?a'/w zhEYx{(58^d_T]?φezqWmU.J#")Iw\1?gE٣쿮wE{pʊќ' /fx)l8_pٱ PH_47 v"靛F00(*նBUTyBko"OѮ6JYYS7n+~YŨqֲ~:݆:F;RPY),v鈩^m{h";F69s9pSDUIr)7mqVWl&LރbK7`+UǚGh]I|NzN!ߞ?=ݷI"]x[)G2HRV]TrxOÛ sk}.Oc3Ё{zN@GE/'{\Dwևٖ_Y :{z9{LAk;,TK? 5st2OI1Elsƅʭ8'ÉlK4Oo'b{z;H1 Oc77IɊʔgy..)dUt6沍L ^I;x`ianM=&}_> [1( ^YoROGD@Yu H5sk-zܻ*dH/D7w+Y[-f) `~|bm1_57_5kOT2QtEV$:/FaVg~hHiirWh$ ɭY1X{Bu8oQu 9z3uh6of37w])Z{RUqjX}|J 혷qMTY#ݪ횎[ܪÂĽN7swim\jw^W+W[|TCsQ9X1dr!E"sAIa"sB*і@5%Y<`Ή.[,X`і"&ZÆ&YOS.`KҊkúYbA!`b ADj`lrdޓ ]< vS/KnKO39|܂bXMD4J (hRc|gv&jgsi`JXH ,Xeț% S Pu.h#pCЅ; ecK(F[LAa<ֆ T]!_i:-|=^ غGQSdA]T`fZ'W5XO*4-iy%UJ GLBe_%c;}U5JUFΌ?N",E.ۇ !pZd6\q}zF+2$F3+\_# 83_YhfNJ7Pt m0g8j86-ca<]ojx6UR{!$Nˏo.↳I7;_T4,w&'AYyI5)y Aҡ vcv;fu/:ԕ~^sߖy]!gp"ܸNV5U90`9_.[{ɱ6$Qs-9˝R$Af]l 5%Y{mZ*IKѦFja">Ն% 1T)b+t@Yr9HysQ b}qs\~1X~ìJ Y h0Ȯ]EfQ-~se՗p.mvMc_dEț&6aT-w*}ڲKM̌n9|Q Fފc9J; i,@iتJu1!-k() LⅩm ߩzO10ƝVfj^GKW,R-x1@עΉڦeN8%HuOҹi:7i F)`"0uB)&J7EO%"dU7V_b0hhw`9)ҠH=CFŽKQ:rńvhG`(3*h2:ڀmT 5T!26Rz|<5Y0{lss)JA]YOeYjFYK'r0/ L\X#r8pf:Y[7 clt:]iD@]|֨n{a';P`ܨ 8L$ŀv2Ppclk&r;DΤNඑ'-D, V:>`H1b8KLG';-#J*AQCbǀ0AimOR';mό.3]'I'2ł-rkjm-seK% \UTy مxn23Ńw޼Dl9}vﲊV&Lе{I !'0!(4xa'lUS[g %B|V̔"~VyD, PuОrZv_blo"<]J[ Z ->^Z-~>Z/oڍ}jc{L,c05 +&wmmHy  Ӿ 8(R$(˶~c[mvbJ6ǐzIzנ|eJk/߲ӟcB@ Ed[4bQI`Vp; f81wM@'EҺ0 ŝ^)mRڿ-matB7lЕՅO|!?4;PkI= OTHKK@29tt(ˡbŻ^HLf9'kquZ|`?FfŮ oo+e ,moTqUoTizc*qCcW;d^78NݙN ,+G!J }%ZKr0G;aHIE07fP1e'\dŊ,@`dR,h00reJH*f^5'-ljmIDIgW5k&~:CHqnvNH3'ޑ7}ӻ4,Jyu7]3WG`{nvphera%',Uv]akh7=D=O&<M{5ou,+YZjz8kOJkl}U K ~ثCx'@Ll'\pzs=)=+k:K.=*W*n}+6puR 9k+yOoG+xofp#OpWAe.uψrۛvĤg!G?~Rqt?]+S(.h*XMf?R/ofo'ixКb/t!HCZb ()*]T|PC'YfOyvyxg9 u\PWG=mס|:tՒ,=*oΈ#(VZ\FB:W$]a]*JpU ߔr^%\9VC}rqifYUKp!z e񇚣<[90zǞC)Ev)?I EC0 }'3-ZHK}gh .<9fd^a[O]ʑ?ުJ:aЖ"U:5 D&! 6\dY}ȎnO!?%eSќ'N-m 'W:ҷl\goRa. zMw>_o#>'NUzc^,Ҫγ*̂WiԒ n&#CtZ}ݢ rAL 0)]W!35 n"#M}|A"j=$gEp*nAX) (Jʬc x:x~>_j͇ 'r劳r.l2o[xJ2AMV\xv2)WhujUץ[̫U ?UL .wnj:ƬsE xPObG= "UOw&׾h ?&ӹPT$APDQBUpA4(@QJQS"[,$Ҕz y[ Xbea 廼pŻdt40d bhF`^^+{J[Mlrj0f_F{]»<Ҟ* &2-+Lφ:MF=#e+LT^xatđ>vU\ݛl*]϶R!5fyɝಾܤrk瀳PfeΊŁ !P@E{ZTS&iJw>XHTQ2JYzsBŜ F! 9"TJuy 90_B|cAh7x2_IOm#]~v;EѼJ&PyARt{;@Ջ׫ջy735V۸i.+]ɖz/#?]%Y9`+wt<1çOs>Dޯks9 Θ}l֋8|å_1ۜn8OlKi@vǸW\(w1OT( bh)!d h3cR&e0ZmVYkk t)袲ׁ kTKBm:q/ x=ݪI|G~]5Y?gkw%h|lvԻgV۷|V꒦2` /ɬ6Z,ymP j 0'4Ca/iJ{J*(.&\196`P`(yղ&ۯK!~b.;q {lSfe~sA;17wM{/)U%2!]ʲJ*yNmq8 ̀/T2"Qd̓K.av 9c Vp5ע`IeYL"ꘓ)(Or@@"-a6Y3ⶆ6;/kU)ynx~,[b;& i?wmu%Yil:sm=3> aޢr`PjLh7,6t4!wJJ(xH3k*сvjZnS[%eIq"S,O!mD@K;kj)ŕ93@z!1}4ewU.X{-=[ -}j7i(ڻ9ѕ#W7>sn~ _50x4!JL3M 23VsLMgmM ݟ-7ySϕir#g6Mڑ ϻ _M̥.֏~K6>|J[ :i%Hᓐ<(VPE1$oFs5!Gqw Bx]{)7},j("zJKGa0i{enB>)I8e5(zL!hehy n} ؋(ծCWGp~Vfc?]\4BZb&,<%ǣFS!--!k qСH/mz{!1lxKK.iU͚Fp6ط ӣ*U`zSżk{SżJL׫W)*5n'q4pdg@`MC1 <Q2^+Z9 #E07fP1e'\dŊ,@`dR,h00reJH*f^5'-ljmIDIWOD<8ys #eS9q6#͜PwxGތMB+mv%8-< >}A8ݭ.ΔR ,@.à$v î+_:ly ]=ݢvCDGiϚ|nu%K:)ՙ[Aar#{647)t6k:"*6 BbSf5^z[ KN9ώ^{O)S$kvRiB{J9 * Le8xnHɃ (!)^F2L"HT6[_Jf)NB9fzrL9gV|:ų]ګ`BSTޟ 1A"9 K^i@UV͎Gm&aʾV`4EF]ɹkW2=OB`G, })f+,Xh(R8 iQXla³bZsjoI~,_`4/%Ŗ\#G@% r!-T夭87&ʫ1TYrT¦.Y2tv2dWqu%ayُq|W(fWPvtaPX*JI&^#=Oi?]r<O$NZdկu"7b2Yd< L aMFD铥UIEHdTUpNƮ b38ecD"0soA樉i HuEEV)!-3$("bs4qc4ZPPiY*h ЧDLZjhgM#b3q8y#-W|ʹX\ԍqQ 8>|d ` JJbu\#W(,즜CNibp,p>F0ϘI D,?{ƍE~!k|:%4W4-C0u%VJ߿}a{mY^R]3 cZŁ3.sWZTŋc4.P!F*:DQƃYbQ&q` i֭[$H"\JlB 1jv`+,<V{kV;#gh9sW'&J.F&Zo{IWUxz0:=lՄ3X|Ucr['Y(>K9N'SC0u b D$&s@Kr7)6Ǧo;"E RrNaQiƝVc  B' ;[A]d(`0 UBd4uEX{heWR T ESh6+FaSa(_\ ݞۿҴ6=1>Pc\F08gθgA3u`Rd"&zԭ m`jWo˽liFvg$!P?𕳅Hr \&ifeB$d f %q(ƅS{vҼ)O$З佽LF]GE8dN&XűH)DJA)E8b 0x~:5@,'l ք4\:y=Q1'_I}I0&1y7Ɇs;.)%ͧzx劾 Ap:(:#%cg)frPJa8[Q]:'7#|,u?Woޔ/^Φge*1 s?Y5T/FvQ6q6-|ha!Ғ$%Ưo鼭 gZ6*@ }JLVpp thޟwÿV զmU T=HHj0TYMK(VVJRy `dm$,$|/Wяk7Mihi*F_]rKWgni|e-ٗC_x쵖?iifUOJW| J/2WWQRe4 Dv羄fj@_uڍWxK|6uL3씕&0lve(IwY2mNY;0A\В$X$R-#4pިޑR&IyI$̱V^)ˁ2>އ.]pj-~M"Lgm<'6 %5~#e[ !yȃ!gZ0!"#8O2mH&gF~^݋"۵1>p XZQ2Hb`F1FiD1nY<@3sj$^&`J# bj2""&ZH0}{mM]%5t\ \qxm~m]9UO- Z [1y;A4tA҇ MJ&11 xf=0A*oRaEPB"VHCiIŖ9(t O*AӒ &S;[/J;7"Po󺁯s9I;.a>2LN SHƼqJ ϵ+J (1t<h2={i=|sW8p0ny}P&'<`!:]d"Nvb< OR׍$?tԣBin;%8|dyDͰ&FhRegtXK!~I8B{L46 })E̮t8$e6umB4w #} O\Zȋ8h(b$"津Jp8<^c hVta馕Ha0QخiSagѐW{`_ǗSU\HNB"=7h=W"MS0DVOi_{uP/{_<;7×*$?He }WA!Mޟhe%28VueiWQZچ Q* [Rƹe̕Ud'Yg/^dJ)X.b3.Tn=9QlC4lRF)V;WWf`1zqrx+@-yJdu,XPec2>D3\_vqѥdCf<׼x4軋o$ŰRqobQJNx(n@ׄԝ>3)ߴo3HUZFxDaPzϯ WZ6k|!R5=>:j['.~[<`]%*0տ81Rol0ЩUCAyUwHqPrP*eȓ( x|:J!}iZtt!\a6S%{|0]|ۇn³&6UCzo6]J^;ԹZz0xTy]It7GZ.u;{ߒv^ o]cZ{MT>sqCC_7XVqg-HDt {z6#׬vjvۭaH-Hb\$B0&2'RmYTSBsǦ{eFۋ0hP'<}FM`P\høYb[F8AKiUg;MoHW[rCzzc_iqPJ''o:Yd5audu:IyRBLVIqFf gϪ*RÓ700oc8M.G0Nǟڋ$܏\οamޏJE䢤L V$$g咽B?zag룷,=kiGĆfjޯ^zT~EhdUOAMD5oF!t6۬T̪L e/'{2b4, 'ONV_zE<~Yk}9'#΍A8g&\#&rTTX)%L"1,D) lS'x^ZI9DŽRn&x92J#"( -J)'RHDcxgnB+Mѿaw"OK׶_[?w׶ö1ui_|-DZÚBEȕzW,Dnc{ ZĄ"LU"X+ +pj*QI \QJ;W@p} NU"W]+D%{zp(bQI3prW*QJۣO8UZ+N5rg*ȮP [ WJpJ0+pc;D.Um=Q \IƨڥXY q'rhWZp{Փ+)$](+MGe,8;2,wdzA.%?>ჟ<9*8gULT^OGM0TEIUL`v\ek"-Q+嶃kR=>ApLD.UWZJv9qG݉`nGQݨ[f;զSI8!);W\w|*Q)p0!.Jr+p|*Q)"\q!N̦A'i 3ve~^ՙ!0|  r:}~Fp%l,dܝ_ @0?}~vH% oNb ϻ΂0C. &cx/ta\YP(9W2:?:jbNR5`0̾C<ϊ>ՙXi ײ &W/!˵%83Dn59<2݋OXikÛ=oދ779j/9v3|%gS 3^L6 "V0#)6gF\Dc~3K~f T,3tn ΜXLjAcAc(W|>e4NU`Fuu#-=ڲܚBۊ\ze3Ӕg)xHnA\\gйt4YNX%Bxk"L)пvΌO/ZZ'k(T)UU{}`pK %[1Cs̃je߻| S͵ՕfsgCV&3.rm^ jKVgJke-#@rY8m0,-rT~\?<4 }d6n0W22Fp3Z a**g(%\ŀpj@ RQ(Æ2xL-I/zށi@p/`q?fRW^]8_r$8 YPJIdx`VX a bs KӵQ[5aVbBlxkhQ\k=q>]<=(Zyaܦ`jl NrǻZjVDLD!e!x%jИ tk1hFs+%ZZE쮌FB6^s-BF0Fyy14)$tZLkQ`wt ^pfH͠cF֑aNJGgABz)ҳ{p Gc`fW o2gZHqe׾dJ)mpf^C6fn3h 'e.#-D;2d{(\]r}ۣ 2U  ieB1qk@Qlz`aZCjۖIYBy&m4p@*XCjMnCs(2!s@jcnU&Iq*XLv qahtplsLNSPeq XP= i7g a5FSbeh%m9 -W%PT8/wgㄺXNUgdA d*WXhJL Pِ&&4jQV]k)d2#"=]hv)0ְcFZ B9\XS`@+f @ZImJNv:PVDC RgNn#:]Lڃh ` eBAo 2Zdp?,p]asWcD:‘gQB@F:|*i5kNor$XޞvY gi&0 ֨( `f)4ͫA)U9u-=Xu2X@KH1&KXbHukmZҘsמျaR^o* 9py%cyA:)R(]&t*T@y!UJ$ѥ5@=N( j=ؑ`5X_߮Ȱn$>^?}Er\O.I"0Fل,ϸ `\ة7B4Z1<*IUk01t׳Pg\cGܠ "epq2csJ6H0s#SH}v=_85'Eiߙ˃>0tD$ƀ?\,Y#9|gP*?C}(*R3GQbTqZibew`93#0ʲB3_ZD^ߟF`JnÈ'9>X2S`d5 L-C`3(`~= D`~:XӦkΆbUq-DL0K1;FfQf%d[pID atbMSlAHҨ56X~@վ];gG&^m!?u7.;z4ay\CD!x ڒYS-ʲZPkr~&Ɗ"kʿakfy].9_Ć 71迧?>nü VTVww?v\Ϡ*:Hfڎ !h;jo9v\i㺲Mz @vh."TBtq[8w\2?~7goпܫ[>)z_{jw︧;/n|=~rgȲӺ-fS[ ym4յe-"VP ^JIk]:{,98szoQ櫤 {I?o2ٱ)ݛysiW> ^6XꩻW =׃ӋKn%Gڽ#ٜTgA_cS$gK#yM+sWi_-V\5+_.׬3;)@kҺFX:m;>RTWOR-5h%epȁU#+=Q'j/Z)):QQ]+9 z+iH] =N ՕXёw 4I퇛{&s={ƍN} `|t~o޽Z5Ȼ@ DyOouɄ)e{"^rkkQhu&gRu)rYx7GK՛ߥ+\Q*n_g/6pd>[SC}S)I^Rl~cjç=YR/arxHږ")xea6Bw )^K;VZx|7*$->^r1 Q5I*xn=ˋwc:>N}~YxZGyv@>6âuDp#]֧.x/֫mw mB7jB?}ERWqfM$&q5I\MjW$&q5I\MjW$&q5I\MjW$&q5I\MjW$&q5I\MjW$&q5I\ˉ%\&q5H\p~5jP^QI\ %./Y5ԫl1m{N6PMn =:r"K`\m`" Q]kS*1!n9غ6;/?Zi!ayHǷn_xkϺ'ܶMG8`ly9|fOzӅ?v<'FmrH7Vmpo;~vnu0Nn>GߎG?=s:q1w9G8rd=6Opog9x>,#uhg<ߛ8qa1{9ȩԼCx>-9=}.7}.8m /F^%7 sGyϽզ'#Ô#q,<屩;cKrLXE.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.A.thWҡZҁp|5.{\]:zH.K~ǣKv|.o6?#5Gip:*U@.7:y  W`U4he#3bgҤauq78&^{28ɺ~{Iߞyr(+I&ˬTsVؑSqj5OeLW^fpo71|!-׋ri|LlMR߮C{Iyl*\AT2qM{[*b὏eS~}o~/滷׎Ɍ;1&\Eߦ:|*jG}uIDߨIMo}D$&7IMo}D$&7IMo}D$&7IMo}D$&7IMo}D$&7IMo}D?zKoR)͂/S)y^w1qz_i^Fw7E]ԯFhxuD)5iE:V5B0ov> ~x?9A._MlPEoVH' `$N? :rV,~,O[~ z3;|~RXj߾r6,n9QJU 8Y\BU\n?\b)2Mt^}WJ{0ʃ#_{y/cH;GBuD=N}OBzkEon?7AF"UL."+U>Ib4?_{wh;>i~'+d|_:g"2jn3)B\9yyr-98S *'f)ʔJDdJņ)'rfKMɚAuWW9bZuym-)@#aHfx{].l+hز'bGj˛ozxyR?;aCKtG*4VR*rҡ#87(ǧTLRiǭq,O>/OWhvQ[xS+)L3_>S 9xυrThR<&pSnRζ'xPL,!Ǜ䙋K|JxdvEDh[T.G3𮼔ލPh !!Tsn)a 1z81* @EI: w g7_AL ?n]I]qPt_S8/ fcxpz+34Nvt &&93L❱OLf4N|Y޲#:Ya+ ~7O%8KwXWWJVt"XU1cx%wU$w䉸V9X=V#h5[֎QCԩDguDIH!.XR 6Rˆi"PnbB'ÐPa >x5FJ& f}(FQe&R'>2Į{) m`&hYpɜlX]q}£*b~R L2urCŠ ->u#B9q Gr,>DV"(כb7/typ+=tz;/o#DÇ:O$(]izwKmpoGck˩fyʞ/#u'jvP;튕f2'bZy7T'lIF•@E搹 i˕21i.(FJ +$p"#{YMtP"-FP#&&,R.)Qj8y*]L:NaiQsa3$82H Z (JKF蠉M8!qI ߓBՍXTo$!#3DD!pЈ5HP W%byK$jЮtbٍR?vx*XL>EDVY-"ne\E/P PwuH& Q#1(Тh %Diq.1KmB,e#@ 5i&4wvp|qq\):Iɡ( "oqŭ %+L1A6' 5D@9K yS6xx \<<,&,0<<-JnnkZ#wZ1sе\Tޏ\Q*Kl**\*q(n+ꅑ<0+PHe *AQ$R:%i-|&֥~-9Q@1)AnQ*eR F %3 MA0 /ZY+nȥaRk(c$b8q{c89"0i5*'>jKS$VΞ 9J'tC[+@h*T2[Iob5 E;G ޜ7Մ7.BFaƄ"$)ɉ ڇDGT 4D@ZsHI;BfݡuaaK\)mk*H єj\p!2- +'1]@ཱྀ>Lnk;zb]Y{FE\Y:#((7*Y" ! N^Y8+׊C?X[^'M3;rDP!`TX!@CdGj;M> B:<;\9bR tZ~Vt~AuasO3#vMkEIgp[[ K:_޻6ߔx"[aƝNe~6v%\ãdj^'ugMq6e3jPt°NِъsÙ>'gгQ\J9Q*x}J儫pnϻrI>bv~Ď>Z8 :CWClpfuzRJ$Oz㳰Mϸ$FŃ+3rIjugfwwS64]}jw(Rfn?o3rfP?L&/\1=%iI3ؓ{n5u#5qFj̳wAa>e$~5^kӳjER+$w,Q|Bnf:`zbu-*_u˝^8?_~~q|ϗ˗xyw/pB4;y~z}uϗ?5'T߼kn]?7MJR|p,nXuǛ/8dgsא:/LԊteNzz6w¬zT߮JDpk1 MvoDŽ7:*x\-jRGvF3%dPݟ˦/ q I H2hh8y+u>y"T }qf !W6,Eae#KH/#iǶ񹪗d!;xD%&%ʴ7.P!r$Ub&XWFZ%C"Pc OQL(/(lLEOtغ>{$P[j_ݡYC;2>[Jɀ)ʉ@`sb8#|`YZYs9@MLP3)I&@8#S%O2#2xY e%4 Q-d85H\\0(9Ddf8-̐ gOotJY?as>z]n>:8t{xs 3wfʑ(qybP̢rHI%82!!URSJ-",ĸNVQʥ(au&ez8+petϷqG T; gr6l#Wyics[̬"rÈj5q3?[a U`"&A99kn+Dth^DM2YA[)<*=-|rP4a(PNƁTF&.,B:u.ѬZ)Ա_||1:4ueWSԄi5'ȼΓ u˛n|UO)*$+!|bINOƿpnͨ~=\g_? 5\&&U79ۜ\+DU|wTbjIejMMoPǐk_CE\^DW^7|/ŝ.7u\?)x3>l;t($|ٹ5'yOS Hg.o>졯\:irHUwR{'0/^7]yy0-{>w&8 ɭ 8}?[էpDH`i vA)t0(c4A;=>.P#PԹ?gge8<7@8f(77c;sp՟i8-q{%ҥk.Ta'ݗM/lW[Ye'=-ђ,5&[ȏ6K{0,|pJDώ]9Ǽܝn+Ǎ o|aͮ%y@dؾ=x7gwj]Mz _?qmsugaA1Gz<훳;>?=;`p*/UYwZ[lAmΕA+-*S %QMe} #03Sf (ܬZTҵ[Y7fUB lF@Elخ3VlF{[$Vyצ>.esf]x{%ZYp<)heWV7~OsAi~{FXOMӄ,,\B yQzSt=On=>O52PvZUM/J NR6夒vɦ jkZ=6eC63r i;,e%TU9r ݍݎ|t.>1tJ DOqպ0 >r9Q鬗ގ$7eoG]'`MyoG4*8ýBqڊf+ĕBrѮ'怜+ȺW`8(6B\#]ԩ(U+8HFWDWDkd"J9* <+GW{ABZ#m(u՗+cS~* 8ǮpCqz"]Qb$t#t-z,x% #]ltE`hK]WDiM uJ1{`+ EWH "+|κ@%%#]06:j2B 5X6"\hN]WH YWѕV[NAgp5h"Jf+'؁d+5&+t9NpX6"\h!v4YWsU Y ƮZ6VǮR+رTHU v걫(ɗ8jFWq&/+Ⱥڻ赆Aʐk5B]T/߿_^0vok7.OuV2| kT՝,*}JEԥh\ߜ,mEb.1T:kJ tJʛwE,VH^{m;k}U£)QZնUFXzEDKAqAv$-???|~kS#K"LD>qV].E r^D2pD~"Jpy g?I žvwh#ՎߔEeBLYh|[TA6q߲w ܔ_vģ|e%oyvh}cax|>xKމWp_#2'`X"C㵁@~r;u{0& w9g5/UAO)e x`b hL=X# s6`M->[y򲩖%V<PȠBޣwmqԁ;8 8h{wx+pF_~l]ulZ]x*`PҊ`+0 tη^vV U´]S5MV9Nߜoٻu=ވl7\w[.R?h?ˮ[Xal.z$_?}J?鮒}_uP)Gt#2kS^9yPO?=UNUi6=}ˇɟ8#xB`Wy."uEVe]PWFSEFB`eL|"\&"ZRRj>QWc+- JEWDQu5C]9D`i]!lf*U(} QW >A Ѿt>uE.j 7tR .]M~8Qf]}A2;9`e5_pUm1TMd02YW?X FW;uJ8Zrf]GW4#]!6ײ6u rt5G]a2KF"\kh_zRIu5C]Z+F"`GW BpʐRg]QW*m#]!0(>cW vyWq.d]PW8H6B\}ԄhӏRˬY4 e+l+M?Bʐ RWI!8EWl]l K~"Qzu5C]#]!d]\tES%XT8kz*WM9VO4!jYWt>hN"`gp+xi몧̺hlw_Ia$,FS^ 0U0H[ 9FӨ+FJFfõ`K'5.6O=k\Db`pbϧ'+ױ$ڐ  YW3xb+OtEhmH]WDZJ鬫XM1A GWd"Jcf+~pt^ hA+4yzus0O3W$$J粮f V y6"FWH Se$;Ⱥ+cѻS!K GkJ 8iDО ą Ѧ?"H6 uƀFW5 iH~zRZu5G]YFB`٬cp+GWD; tL%\æ3H.y] G|t孲2>C]mH~)ȳ樫`: pѦ]J̖;_ŽEN . Nh;uJ(\:a~]EQZBBվE/ Gh,].(."ک]Qzu5C])]y.ltE|tE/ +̺@JW 0|+µEWD;fQ2H@1FWF*1u8Z#RQZu5C]YiQs+|+ĕqǮr3j>rkͩ3HOtErd QW^I#]!VW$?Nd]PWAio#]!p|Î8\fhuQB+l;,<vgdĝxN$Dp(Aj}l>E]Ax걫H\hK]WD| uB2O9]D(HlUՑteVD.[H:|j-m-K" L;ĵ9s y6rۯĕyc'Eq41'u Qpmw 3b1tR86"܉6EuJ`P]!|+u"p?`ebK+ JkNC䣫2jr =#]ltFB҈ʃ 6|tEq҂J~QB|;G]#@s2HP;zEWHM]WDڿ ]^ LZ$:]E1QtG)ddվE/ ltEhO]WD "jR8Wm? XJ_~l>4IӖ:+uuA)0Ah]i:FpbBJ,jED .u͒di&{Nɓ E*o i4(UQWWsx#q@Ek+L,vqte t>tEArJ)SQ*u5C]Y+|+ p;Diuy-GltEoI;J8JǮf a0~|KŮ _޵⪼^QO'7^},gWfuvg񲸻K᫺?YS7.h+)Je(Sl{TfˋLéq ߖ˫WoכI3 ѧ7zF'6V 3wbkySR- Ļ k0Ҳ-)lc:e,RJ鼂u'h O!VӪ}u5O6Z Ox{{po%k%V>{i7~  ^Tn{7'yuS"EIɇgJxܸ_.= xd_ $H{ GZ#)jic }Qkqϙò*U+6A~lh?8?sϟ8EW`qF7$Z.jW#膃Tnnj8X ze=mXș4w/Tl}{i2Yǥw9{Qp–`'hV[WL{zl+۟햇hbmc=0ޢd0ŷ)y"8L qB |}Ǥ:PXbPGomTDɨh:su{*e^}Vg}#?t1,.45%X\*|1\m_bTH i aeY4K^ V-U uA)%í^y4\tl:_K#_r9ףҧ^mPե~޴0~Q^Gn@۽q@ͧzp1%xWv_ H fW'koMRV"z(?MʳS ?j3(0ٷYlX֥]tzM4'IYR>&lM[> f ªR+:︄ʂ7T3)^ %H^{Oj,QЄkg ª1&+,FK{/ڳ0Fǚè3ozۇJ!қ@>c^et֠#;"A<m2yƒqThD,W;U(d+ zAyo@Y"1Z(O!ҞFxpblރjaRkj]vQM7pqa/U8O*5ѠWEOεGX@_#ӁW ΋lL2 z=*k]wU+#̲ve tXtє' )c.t\X>,sT t]YT]ýY Ē,zC{M—y#$Jx|@nmNoǣtoR ,fڡ`Ƕ:LsU7LtX,Sl9l=\p.*sB:t]բVVir;+_ƄIlkz{Fnd!:sqsAeDöyr  -lJ RX*8QΡ; n,6hJ{;Hٻ&y% Za@2NPͽ$q*$;($*{gyq&LGpGm`*Qfs3DbKĶ[SH>.s`A'OW'1/F;jwo?X6mۿđkI8#ί}ә:V%##''@MD LNTUQhH$ҘO{YCsTMHΔt }Sg GXNܚ:7;u %"b39sDz|ǵB7E5] El|Ү1tT 8q⮒-qjDZàMNg\ -PԐ U,GJ9F+^xRbGQ;QGJQB@,fC`Jx&5;Y.Ý{fA -LnmrF u`>RleSxȎr)C'DpZFs3]π'+hXkmቖ9 Ȗ; {@v,ޏ>^U2{.1T-9n7n6 _b(k1PީBTk 6E`،a )m;5VfіVv6nn:*Y=eFIr1i QDJ#͋żH5>]n59r :-صt`hco%F8D@ܤYR/| 2Ĺ&X- 5d9`+Xl>Jk\o]ܽ;o~wj^w>5,;$6ξ+w#i5t=Ң)`Vk.= #)i @E*D "1<^ǂFzK[mSX@ -eNj QG!ЧSR b2$hb+7Z3X[sL2a]'+5[6no1pDA=iί\_b}28w`贊{M{M{MѾ&u09(SHk%LtL$L̿@Yx]°[Ј*tpTz',v}=$9gUVdً[@^g5e9s0&teb(hEMa1`SЪKeQ`)U: gf85c2(ЩR enM=x/z~"H m{OֿWn9pWt򡘥a)m֏ stBBux2⢶@ɧc)*3Wfewয়e pxi=hⅈDXH-"%#wzkHRmRog&jSKzVWH\ǖsiKѺOe2vlÎ5s &eJ,(#nH)xPLlÎ$Oܜv:>h; @FKmO>,w_  AǜP:%i-Dg_bw=Q8?=TqAR0c'h[R)!6*f(y\h d.F[4LjMeD;Npolg<T&0-}kɝvN 8͚ #|9OCS~uq{]bmϷn]z0pX)1E@¥ M*H."S 1D3BIC M>+:^X3޸iLkS0]e N$c hJ5z8tP\fBVPK2|`'a[z3@GueF%K$40+<5'@r:/F MgϞR`(TH)U*/dh`h2D F}]G͑3=íg:C֡3f ~A ?a\$[f$_Ή|Yo|Y cIi.|̅;/g9 S.~tYݩO =0VUJljxۡ9V  5Hrz0 ;ZqNy8qe ~xuRpfY}Jջu8F=e;We/a~vB(*tj%=  Wү7pIЖaSm!7PZ,[iFJCu|_^xYgוQB͞OQ>=gk"4OYwua &hcMB鼩T4VS7Fj̽#x4Zutsp>57NOcuȦZڪK*<$W,|>,z^z9#~JOho*B?^Lhz߿_|=e˿|u=04P۪W)W(pUsB۪[U͍8zM~DzϺފ:VOO/FAv$Wٟ<鴉c!OQWPmtz"*|3"X !`@S^<w zUPsiK߲]̖T1N'H~:<fW` w?'g$<eЖ1Y?uex7ؤ u'B {,ȐCae#KT])΂D&AfRMC:^d{:Ƽ.zя.o88yBEiJ)_թ/؋Y.l_D/I+-]|7K3.)և*jЩ$SΩ!Dm BaP}~+^Tz\nB/A6XF9̠5YeU9gs‚ ާH A2E阪L~~sY};m崾G-+J41EKH-ȬR1hb2T h89v6D(|?˵h Gsܦ48S߅qɍ.cU`xZ>P|/@Y򥔗*|8XR nZ5}Y3#N$3rvB@1k}5?cRVYR5:PQ82mR"E->AX#&Y/ݠoA_4Ӌ1]pA9E$']*8cC&VK&,@nďLpO,gV+zNAsZuһVkO6 :~?0H<J5c2 ޽dz^N9m,6 >|FѰK%U,^e_!]SYsσ9/_gwC:}^g9AepDW;~נ8Ϧޗ3M~y?,m,bYք;$rS B{8FrIfo@i(y#bY-_ Gg'>J̚oE O)uck}>>.aFY4l8/ pL@o ݨ媛H\ӺL9w_H5n&7޿snNk9K'\2N|LC/`qxyIo' /T^vy@7~1 IVe;eTZ*o_h1Wt)|lU%:v1.$ez1J 颎G!Xr$S{{=R'2ktqsS `A';M^(A@jr9OXeb-Ex"$3 `(CzC?ydUNPț!7k*MO(m5OΉ~9+r+i37 Y`tp3:\[7mo4\ѸJCZ&(H\gb]oo q=7H~(=[G~& BbY^uޘ֬e=R+-*`[ ımj| SDV+SRBNd&leVA0ɔLEC$CK!smk&g7wKt877r5-n]{vt vVB=h´LGQf䱇z,-}+ $};d!A{FekUVǛ*|߽}?r8eԥTap/eɭ =@̅A$2qK8E=簮~>;;}w,2NCMU4YCKMN2R'X.V:cST mv!ǫ%hUsn"av\`۽<+?m>OW0hR&PK7i@Bx)ќ CS7hzwaQr|+!nHd,Y]HlfΫd;vR%L\S#ϻ`.-Է?={[S|5`(zZ+3 4 }!,.{{!.jZxM!)Uv{Yo{@톗WEz 辤I:`DӍV67hXnZz\olܭGpeDpԙchC}mk3 hQrҎ"WKQ}ݧtнBԹlRk/i7S˫6DM~b@qu(f6oW2@u S,DƉd8|VeƎG⎥sfC{#Y.޹ s$a1nKKSG&z]Jp1MHd.s=:Nym}Yg<0y#dn#C}xxî*iE%{~Qst>@enui&Tʤ@.$8Jg`QAC~EFaRk(c$bq{cϹIqQ[%oߛ8[zk|_<3zd,/$PZ]9t;ǛYowlW):1ކP^t j \|UncrQY1JEdJr"xpB!0ёhxW10i-y Bd, 'Bk4g83ͷBJPq?dU*Ƴx߉k_cʤڽ^nJ3sɯ=&;0 A.?q7[fўyO*Y$8Ē⯩.|̅t;̻0?nnOݧ30V⽽.%&'}q͜;`HP@ޞ" kZqNy8paƶlE -'Ф ԁp..'\ōW!p$w^|UPD/ONWrBQSAئ6~xF]ZqkΎT+oԔ.y~8?n>xLn R =qF4}8{E.mM8>vy׎aRђf%!oo餫Ռt6*?40V={1٢$l/j׻jEBf >\ +yG=Ŵ>bU5*ԉ_FqFǨyG߾ͷϯ~}ѫo>룟߼U/rj.zsW~CM[7 ilmEߢ]I|voNE[0_Y O&,dū:ɇN5Q&]'WPl'ujѬMT.|$VbAFͧqw#V H\nv=R\lLӔ ATf0lO[2r T`7;`Ix8 $be$peܬ?sӿN6|L&Pmb0H}Ğ 2#ͳfyo,q#`d0{w粚ݧe#.ITRk1^N{bcKB[p! /N7,ޣ٢BzߎWNǫk/nဃՓ[ Տ{8]O`M\h7}1GuAVp/q9Qؿ?~2Ҷ8z~Nϭjf$g)I&@8#S%O2#$.=Df5Y3*b i!ĩA+%S2 @MDC=FՙwąU(SOIΏȯO~Xgj˭=8'/lmn[k[>kp#^3RD8IBb1(f#Ĝ{!Q%54ja@Aˍ06icB*0J5Sb"%S/rNN56^< ;1i>>C&T^[Xb}5M*i48Y-b*05G"`4IDr,ƛdm$LΏ=3 gg EXJEH/Zt`(Aed2"40-].0I d8=g#dzd XK6+.ƙ($}jecAHBY۬߿˻4ɺJ-:&DF~X0/y7nV4|CV"DN6!!FK9GCdB11ISJpT&;: h*tJ{"(@k ƜШ+d"c#h\¡嗷V?@5?{;dߒD'((GuPRԎhU9\6N-]ĝjʣKgs{'h6Tr?߻Ӫj~,A }nivu^& wa5a~2d8)ǹfʐ/TӒ%I81>).Ƈ7gIx\ǣG7FQ5fqoKS{H-̲ki]&Br.nf#gpj:o|q<\ŷ7Xrӷս/=^e ҄㦺?'BP@<ȟť4/}첓m?ʉvv>==Sa\qotW^֯[|bKMPFt˵mpbx1k} _f̝Y=:iphtԅjnCt1 t{AwJݧm ߍkMs(M:~ol]7GHDcfHX \v]#Qc.\;6ԹZ޴UӳTm >HMvPG56ʰLn65umVSDd3w4ԚMUw֝~<,D .ʮ(MϵKiu`\-4em $`rIeY/M0*%nH/ڳ=៟NOGX=  1&k$g$87,Bʼn4y zYCPU؂tc9zInv#/{zmJ9[C' Ъ[uM6tzLA)U1&+,F?@pO l.=Pw{2:kp#;"R!tjG CXvh3΃ ޹+bA } {KʐcF)RYLl{g[zMҹ=rP3K`7cUg2f޾Q.\+Zٕ+O]&BJ>Cϣʴ{tYgÇ޼qưo]׽7j^㴳yz}8N>|2:?"._PqtpRi6wDfmmvuoMkƧgC[YnmD9H+XTJ-lJ >!±Trh)Y.ڠ)a|}z$o¸vA+\Ts/r I1ƴ4Xb_3\SwLw;C^o ٮB3{]+/`gW2||6*XHtj3c|0ao)P=oc6Hl`5,zGXTN8e%#pzfMе ?t8\KkxAKg κM8g>]%0]!1_}"VwJQMvP)@EQSd !IebJI, ưݘ 7dudئZr]yldy$}PC^qcР$GcBM9{[1 B2*Z驎75j!o27cX{^OC&9 a#rPNp{}Azl~|L/'NExt]/:iJijgFeόp `M,5uf)(Q*CeF:f1t G9kHAH[~Շ8L z%PJ-)d>'O`NypNzWz<+;\VJg)Ȍ"锨G#,*68{N WJ.*yqm}1r0ۖY+@YTTٻ޸#W~JՏaXDH+7/~F%k=53P/IW6,y:unFnV,'.^z|0s3pF `زjlGoSRŗuʲNyYk\|Nww}*8mߟޜzՒ_ VܛZɪh>JŮF#% 0Ӹ^]SEuTNe+Nx0_)&˿Lz 駈e^YX߲:=qJs^b?תX?XlO6uX?wbyxݼZn)zχY{mH+l8;[?zUvog^yg_p6Qg{}x㣓5 ߯=Ү6f޸qx{dpxo8cwz|a rvvB];Wd֋5-Pm~ zD\DGtW^-rGtW[sYeS;s9mѦfetsnkjhkoZgZ׉UV_mԇRnN4`Wm ]OVr'Ezb2opUBU7bMm6(8K>gngo}ywYPDۄ6ѯ#$K="-\gwGk~="XC菗ޑQ&uE^M"QhE q Zy|0,q0p7y.ھT0r%C̔7L=d%3Īq#P}GI=#lqqtԜe)ڗD*^qc UgKx*߼)Kv`Yx%k5o5̥Ⱥ*-QMAy sS鬜%QKysaLɅtVT\U)6 fjHiA5ሗ/̝SHՃyV&9Z m)YRG˷olS4sc}Niqϭdlr=Pr]SJp- wϭ—n2y`4c7cW}RёJUF-9+Z=L5/uSK-m*F)"*=`i4DY2dClt)SQ\ &ϛsuL4ƈa4BqU7*)$/{pè+K*Am>{kU 1WmNz֩ P2|,RgxB9i横:^yr}4pNuulaST{HdtÄFNqNwr26z=򰡥"7g ,/R>Ic `֑Jb=w|֑]Zy;a:ir@,x&TPd|" R(- O9̀E&; ӌ#:CI0D0QTzs`yUo<X yql&_˔e qah Y\@f4giFm*A~4V4zdF)7@ 5hSww m {fAydS딱\baNcὀḈ:xky Ӥ22)L$(\Ҝα1ӕjA ~Юd!_t#fP57]Ґ?;1m2ߺDa4ւH@ JLhW4 ~DUuASPbtX,p`&Ď\j0RPkJ v(R!JAT:|7 d&/ sAGM-VD\ ̑&EUW4Ȓ /Ez@?Piu=ĭ[UW Yv«qݳ. P^ϼ(`Wz,I KV#5:I%5sA e >P@#qҰI&}EXB7b!z_4yȠΌ>tc_X{>~J1,8WU)*S;( \^g3s;q&b<{&[*uUw FfB$a1=w .M Xٸ+MT(];-6*h`efPDd´.YE ]ʃJ I"922_j)Ca8ӥic8%f A 9`ɠQgk#حAFC;T*"O.XPD)4lnQ}\F9KPao ֛]e؀yߺ|n7=]zr?.ViOGs=W&YP@0u0 glhfѳq9ev] MggUCŬ]ǰj^kIQ3󬑲T1-6nSdlֳҌ9N0Et%\!mE:B R K;hLyxPuU߼Yah*4Os+Ikj]Є:P' 9B|G?k7 /oV2Z8R.|.*j#-Nu#z%Ϭ=(W11QFǤz58 pc[M35Pc͚ Rת&_fJ=Ld2Վ *WવZ׶䠧G[d8Nҳ֞s66gs˥Z`Uj&"˱XTZIކEH 0iBeN V3&!KkB}w;WXx;j BI>z5cf/~q4۵eሗU7 ?]{o'u{u" )EJޝ#\kvFī҃.JDK>}]:q fg &3p֚d+ \=BJu7UkwO ggO?xe;P-?- hzwqwы J-2UH- ruO޷|ۿ\łn/^Xe߽mhXҗg[juɩaSBaηj T(|[ŚiZJYh鷌0^Ce]99?տ.a?۟맟.kIvdklb?ϧ!#yp&l = =@*_]~V_r;`SЮ0bw0KڒO*׆shwރ}V {`*}t?2D5ON~:|}&oq{NW98xC?xu{FqAJ9g-Ǯ @ &^~]Pvw3:& kj~pٻ6d>xd7 v YAӚ"=%y69;Wu_Re<&urڸ+s?0d#xCN̍AVݩfa~LLX#2Vl,\GorbSڄ`Fiw`Vv4 e2x-T{݇M/f"nsYc͊FSoizi/VZ/5Q R$FEmD<$2` `.\pLT>yYLfi'N"^(J:+ert~I~ Eո(OOЌjv߇tEs9Aoz~Lc/u'P[,mЃr{^]A eK>_T r5O^  q񼻲̲p'j)fC `q;8RD2O>T,Ƨ2Cbˀ08L/h&atͬL]L C`r%/0|F^XHE b:0NҌv!n؉'Dv6| mgTTKWf/H`<3WYiEǸ],`3,:$0Aի_dGACc٪BٛRL.Wk]MX5.,~qU^KXP;uC6KlJYp$*8ʓ+,=Cp2cĜbRL#,8ZI}-\Nu`fg`*7!^^k7|!mi%A^~g҉y g9tV \>_wK}pۯA5i7Zۀ$C[OVe" fU&&-? G8 ?/?"&/]L*eP}R4lMi0rNK$vnz'zsM~:0]L4[sM­s\#׾W?#~P0FUt(XVHʖ Jz;ϤG|]>ŝ;vCr.͖m6Jl"0uB)2iJ*SS>}>XKE 4G1.₋[E|8oI]vi+fʺZL'W3#|q}9a F ƹv׼;33h$]MS:틝F@5;)Ҡ* (zKIt&26Mg7oYπG$,=Y:c\.<!D00G\y)nF !rhf `hf)1HV% u/^Yʓck|$e һత$5()T[(R"8 ic QDIWc @& Կqdi#{"Y\8/ԡm|иnL7TY@1 ˂.%:'>Z1Ž}|nas7v|v3$5i"@J.A(6 L_aLYWO>u=`&,_RCv YF.d]2=l[uK$G32qexe&ЖLY̒gu>ƙA[ޢ=Zi 2]1!SaA I>H!rIiO7)ApE?9tvۛXǍܮFyh9Z \KkcyA+v>^v-:::S|ڴhFS蓾m!˵%dH]6$w7eПM|w~ma;>v%.~3O++$;s4g]=)I ?pV\F (UU;edfŃUūepvw2 k!^Y;d(B8WW?^I\;G ÙV,`CBF]L'8b:/ßQ >lu֭U 4N a!i`(;d?g3(o ֤X))j:@q+ ǿw~\b._߫7߿SoRAv)E &pgq)M ͆1 c\\r͸wg>nU&~~ DE釗}m>v)Y']:-LЊ l~]fWuQҜ.$Q uvᢢCyYت#b YllG (ne#2󃊒pFk#4JFI (#&(mt(e;k\ }[|+ǔmiv-yQn9xM[\(9b4whOF>U0Hrμ@ [mh%Q$B(V0TH^0E0@iXlXbɽA+Mҷ>DG𛚒/)Tz7c@jnsoۻfS |w\8|d"mSEp$Hxe^YUmV4#xɕoEOi%ϳӅ14e`{#0ŽpSRHq Tl 5#y\7ZSԏD- K'$RB]>BGƸ>&/uZ ЮI 5va{L!ԄO \5<>Lwk" Q қI"hN,T< hA1F,a?Fx x:FB-bN[_1SW#JBcpwEsw "4~y~2rP~x[zKvg[V׫0 o?WtѹNs_b 2*D(R6 ?'yv:*ֿyW+Sr56y2T^,=wsYv--ʹ"2ۥcU]U'S2H\+:͢SJeRt0h(jc z//>1tJ<; $p*ET *= }G<N)}ܤSDt;{~Ρ5==DZL={ڳ={a:9Ǽzٰg! ,!sVQ;/DśʜɘW?&kxnϫؖ-e['YǓp7OjMAR#(^Y9fˌ*cZNl{Ь?jtFQ~S&֪ᑧqf,fV{3hcުtK2Q㢋dM홏[Yo.M;Z֗liM[cjfgsߝo?xFNeSX0rHb\$u\PRȜPJie!PM EV*-"s[{#wG$JhC x{}qg&YOSq$B. \n n/$0ZR_HI$e)-q35=O?U]/dӞ?ykųZ'hP -rξ ACh:8] %1^iIUiҾDE1fX 7OUP fm`%Hb%IŦVi-A)(ZG[)X N ;R-kP5m&vvݭ >ED?Vsm~zTM9r~8MVY޻c/D1=ZiSYS?]3]Y}i7N?\4`.&.NZ֟2H2F$Vαu Ofl&LK?0j_6Bc7Bl_N=k"_ >>Xju%pW7i ^~Lkca2LVRYEA6%3L<9PbN.M8[%(8Oz(y!ʮ  Tkdl&vdlUaa3x^WNj~k_^~xN`Qj.ϦvB_N8bC ELM)h 3 d0iK:Yc\NhljFr C5Bb60qlg[ض@OMITDVG+q#v%s_Pvj?~ހگI tҞH$k(mFjLŰ&eh ):--dfL$%E&H\2<+"QM6xLx@ 㱈Dq#Ь gl!g!bVޤlJ11s$FlI!EI0P"31 iؒ*ٙ PB!2.N)l%Elj7VJf&TVփezT]%Z!@̛rYB%Qpq/xL;1 lK@Fn6:(h я-ձ^Kh&tyڠAĄ4j PyKAC6c!-IgD繧#e,U<Ȯ6X1圔5*&;/< dDFrT^Gis݉pq-Wͫ\]~F7}oES_ϫb-˗]ǣy{^׽ٻR*W}]Np%Yh\0R3 q YcwҮj^u2:KyR:PMJRv8|HdIoL(:_R)8U65.!XԵ.f Ee)@). zo&I ,xOOx>5drfuma[ս9h.3f>+k:4mngJ({SJ@," @ % v3uyZBY7J5h:uqu4plח]у ՛{|7Հl'$7@}?yK Goy KGSȟ5Lsᡨmh>'WO 㘧ǟjAz֙gߟ~'I렽'cS TMJEFuUR<8,SQJ}J`0.9Br3lTbrEKlçg 9[,uI:bKFP5<"f sL5Ae֙OnqpT\~X Kϳ|x[ޓ~ަrg'gY f*)~z^^zqF]/R"$K.y$zLmCex"5{Z|f/AT+G UY= _wL>vHk\oG̏-nxQY|:noa4sR}DbފX "94\z13q2/Pr~ًd^d^VC1;\U)@uWq@pUv`ŵB \Uip J N\UE֫}+X\%]UiQ;\U)zp␌*9Zw(pUݵcpR:\B23> bA1Xc Vi,nWWH{@pUjW,PJvpFrR襍ڡ]ɴ%>+׀ɺdwD>Ŕ;S^SGP.`LA!ۉg49cV'yXR~}FQGVVY}RhDJ֥ZhJR:((&ZR$ :/w6Kƒ|J%$ >8mzusM?^\ow">Z ϞI9%h釿ǿώc7.eb:֛/{}"=zNGps/!xAwV:{%O&ePѨH : tF y=e >xٜ(I % !:Mb;4cV[2G]*Y3&Nj~7Q]b![MԾ9ӳ$?IdKy,6fVR閈Hl3LPV ELy>7Rd4}5%Oy ~Z~ŮT;/K1d.ǓӋ˪=.Y ٹТ:lyӁs bCNy"/=~jmլ6$ޡUzgj6le0Er9 bdTB2Aې](Txg#5%FJk J鐔E:b4A DQz\*,?`&FNn0$VF86>x7eͼx@ <l$Qa{k9ٺ"2!5`@*lgwJ tվ %'@6Z!ޫH:Έkh84J(q }oq&./z!{':bݒ>uZlxZ٠ 0rClwL4:M|`[.hZuIldB2DX"OH>z&RRIEyaqPؑw-_=57AXt=EٝO˧BΧ^iU+ŻYhۗtHD#)V0]2w„( bC;3Է[칅ju*nڭjW}\^ΉZM[[mp\}s\U9ץ n@OmLf =Ũi ك1*' + TQ%Oz|?:x`¼ c9J'Kf0<]Ņ)\Unߓz(M*^4,麤dnOw:SrIQU"(;CdKXT[v q;O=V9#Dvt/ƧH qE#-E[ p5\+$BMA c1).>$4tKQvָ D@d)|\pQ)M&ho*zYo)6灮k=ߗ |}I4uOu]>lov-V=H+_zQJekYͩ̋%OeeQRhVBƔDvQ H@-6瘒͗t?} Pۆ7} !V:8:yg<#(D t$kcP1ӌ.e&+}C6".'Yɸ$3 "69`\!J$LEԿV3L$Y=LBRFFxQgd 'R 4%M_*1\׋x@NӻyۻGt~9g~Y(KpF(Ⱦ]LMGtZ8<7\\c&g^+>BgZ{8_!efi dn `? ,!)ۚ)JMRTQYboW=un},T=o0DmA?W aZYk狯ߏYٓ.Nsh_}XȗkfSA\|F5?kΆAhTV%"XCA vDNتF%.p]%[n~怟t9 ϐ?'@.?:7ײ~FWvHd^힇PfԚh8q8˙aW!p`IHԌ JPEf08\1RlN\!mdYqP3XGQm0>S P _Aty Qػ>{5\ߡHv.Ƿ(Gپoxr^(CA&i&IrhZ$K )Oڧ@B,BY|ظGytLF~sߜ/hy0 Z YomTgsY$$bCJb2 ɁQSJzzD8"乼C  D F(RH@L@:{2O82: ÷[e/9T~gn^؁-6w}V ̓\ R&b)"`4IDĈF/&l =t~{鄞y-MbA *q!j)qɄEh`2ڜ%ue'uFy\i1@ pzdxZKVP@)k8d1Dz19JVV8$bZ|pǤ^:+(A2&ԝHQin AB|gn/TIg!H+BMHђyN\<јY)%A8 =q*1NfmYxIaAA <??ӗNP=k9_ Rfjœ_$D׃/?٫lMe~Տi!ma!\8I&a\wJ!:;7آr ;^LזÅFΎ4 xMՍZAڇ>uyF )rp^ϸ|r~C"F5pnm=孟r{\V{KmWm12w8mS]d}w:1TZ"}HTfU3T\%b:]ݶP M;nnz-І"Іm )txӳkyf+9~HrC͑9r6{HmC>Z G$jMz-ͽclKI|:ܢu"¥cVۧU*-a=f~0k]dSz.7ha7)bg!Jd;5Դ#X9 OӨ?.j+%xGc.$D2%a\kg3N6Qz¨,;.E6gG!ğPcGw1!#Ycq#Tba"-*N)W cuT>kM5gЂX#w(f(I}dDF[JQ3gkY3_XxhM$19B+ᒮ,b$#$Ψ>>rJ?,ロK:/gz'"9R'ژ HZ$p_r))\H>WIPDRBHZcLH{344-&vv]]CgR>kgUNЄ1^0iiI>zHIiȡ7T.6ǍSܬfXq9sDWBhSy"Z $Y:kuDXX瞲gIsf Cl$)T5H*bz@R 0\}e(!jRcp[*P 7Z(OJg0,&Ύ9(6B13 _qJ}S ƒ4WLEsxe⠂AXr)T_K!RUTD[KaPl %ͿEp,£h%7-5/ךPXQ]Qc]M"ñ^Var~y,!ɾ!HBe#wq6;᜜R֭]n{ifq~SɖSDZ4P$)z9 {_HNFF$y4e*,{bE  '׆6RyDp %XLXb3c/,JQ mw־r_dfq\]|4AFɗp6p$ByTPR-Ƞ21DTf6:I0dcEF \預8:mGMLXVmJug;b8ǂfQɓzV, 97N ҂V2C%#t&FHϤt7wJ-ddH(& 3AVBژky$I@s86͏}FD#b99K<1ʱcyg᝷H& Q#1(Тh %Di1HM8xԂ3>`(PTs0!%̈́|(]N8)E1u%(7D&2Ÿry?yM!Ybțbek.Caٱ/x(@ؖ&@Fn |Oձqwp#3Ž'\kkEKBQͧTF-}Jp+1,`% os,Dz.tztk)FA xCW+DsJWL4X є omTDɨh:;]s^Omʡtm~ ˤd+= F qtrtsT.~q T]ftݮ»w]nt)FњYe"='3{b!57%]'ͭu촻%LyQjF%24/z^iy>L' -nfNԳ|Kd|)W:ۦM7w~=)vM[Պr R[nmH"^U& X9r'uōeM [~ֻ&y% Zd{I UH MxG3T6 yT9.E}95>K 9+\_2v&_[nShٹ}g'QFL}Ǚ:V%##''@MD L/]|núFeuϠ7T~yTmDZ]5R]7U$,`)8*g9CeI'-!&jo&ꑚ! !0%<Ý FeV.Ý{fA"-Y٪69f#J{-|2@ ˦| :2C҄SV2tK)qi:?;_PÕ~rzyAKgS9/2QOd^'0We{H|UYcj{}Y2iL i d8}q]7(M7/~O޹C1ď~>k"eלb W_~jsXi!襍sԋZ2kwx\Pի5q%qAU zp54mz8vm\4|5,-"2׽Z[>/GԾ_ /jquЄwx_~4TSh^ Ċbz%tqy:h_^{Wo(Q)s%˾O֎*ۓAP)CS**-mN6ՠ:A+Gl+ ,<t_YwGJIS1{șP&`{ʂ#F) b@ 8R:F"P+ԍD1ncS`^^gC[E4g.%/$^dsM.Z01@UU-bsb4,Yy4, ~rvqC]#z="?Z,Nd#mQ}2 ?/@4P*JLVk0?9n&r8PBKk p99QG!"tJT dLFXBSs3'I{I}nפ&2s1cɊjFVꕤzbGK>oK?S1zR:VWLrV^mW[v^mZg  2A#t Is$= ]koIv+>eqխ[kAf1@/ zڊ5$5seE(SuOSk#-aVeU7]MW W{>EAbjQ=tmaD'<ұc )tPف,eQ oHVHiGxL&) >+.ճ Ep%{i-&8L%tި^@דSʖɶkaE3;|,w9K|:gi }>vϻqhzCy-4G,4a顯^:EzDW]!BW-=]U(ztZ wNn{W}3?Jayy¾&j.1[~1S/OUU1㞸bwDž[Cӯt#jw5&_kH%$W`;b%D燋ǫɄZL,aaUB L\w{%R:*!WQ-Z1-vJYM4оQ>0M muAFiGN-:ͪ`RR*ԩuӐO" c, 3N!koyEG-e5U B}_~bU^O03z8(g{Pg7•/Ug+J4_'ֲ7tUh=w(($plloµQWS=u<]UD]BҤ=+R*\}@W 9}2]1\%/tUѢ8w( t y%`UCWnƮ*:zq(`_#]@~=+DƮ*\cBWv=Gk=@W}+$OW,O2`7]=zfzZLfq(홍]GЕW/ K=+) ]1\|G5Ν]F& = ظU{ݬuž;]UztݫInRH6Ʒyu|ډN.%F[\jEh(Q׹{# urAo@\RH7XEV?$p}!يV&(H,%zDW/x\ ]1Z%~ĭ\Еc]U'"슮*UE zte= G#nOvx(0A*KOteh{CWq}]JrN _߷𴹞.&~:͒//kJ !泵ˣ1y1= `t^gQ%"I)-o?ӞHGi2ިk/4+" _""e^/L^>T}n5 ~U+ZH:eƞC h0%ZpBԑɌRDR[4:Gk_?o7ߤ$E6-f6\zwy<8 fp7Wx}sE7ƾO~g̸ϡCfƗ}ߜT +V0 ɑ,Elb JFRXI). ̴M;{M6^Fs]JddG&#{,144Cj:K2[w'cLE4T!%e4Xik<9tV$tD%^XG %yI\db-;Mkr6.8I9R}$"R"[*I<)$3_B K78jKi4=;nMێog˷(p5JٗM`t2\t/zm~<*7{(i~Rou55e.Ήn.isDdH$~Dsvh9?77DYL;Z4ۈ}:n&~1{B(ads{m$E+}pgѺrFcpr] z* `Zx܆/Lxn kygulLa#$r$op}2MޡNUDLޡXۍPamj9HtZS5U!ȤuAaB4>hMIJ"eIhüS̋z. ]wN 6:]r #C(ܔ(e.(嬋e9$gȈۺGkӾ{Ħw]qMu[*]gs5Xw5lw봶t="ڱ]. S6SG*D)6HJ5hJjkD"$g-8F|6:4ԅe^UܓCs9RAQ6FrY9ѩ6h8euߤ2 A26h Uڨq738 ʡL /WoR+qoZkGnt%|Ŏ} E/`_n.vY^8f-; D93PHPXo!r;ȗC>E[WKX2fn0pB5$Bk$UiyQܐtX$[sڛP&F+.g Y O8 Ξqs 3{XA{|鎆ʗOK@h.@fٱEXfO $&B})A{3j")CRC4 xTe>znDAg :ZzIOZyOy6 )dP-֓$]67d/U8j6鲕~X5"bL()攄B>j- Q 9s,AHyFDۧin1~ѳs Κȡr>*6dA؄k !XN CPJR_j \4aXAigrht莣CAOڱF4ioY֣B.>Y.H+.y*!P\Wh*Qt=G#xzq\^mck<|&gr [|rs .8>_qhI?-UfcPZqfEfbd[.B&_&ge5Ko/_ڍ[f^ ٷ6E^nQSz,?,ݰ2w-7,Яbvb^ŕ_,_|?si:-9Ub] y\S?^pk24]Ae]}˼I14cdf55gk~.yH"'#S z4߬j['q5fG͞`ۃΑ((B$sxm}r I)IRZKPJ$Z )X s#BQgpF߲($n0J;+̸eN7#K%Ư%Nݛ 癤%Β?8[ r?6LM9tq -F)FӐOҗ EmaߐV>Շ,YsRF!9PO"(wmm$~Ir~`sldwb !Z"ʲ6?3CDQ%]tMOUu1Vz#0Y$a'RVw Fӝ,Yh/yn@ehnɴE%GچJY*mi'ATJ"+|JF'ﵭaOѓNhȨQPA(j9E6yyr-p% )Ԝ&]\WODnS3=Qz&'zo؂)r>m{vn=B$[.-k{: h] w:DIWqdJD*oi "zuhK[K=Nˢ4XIۮj[}]Ή|࢙78Ȯ9yZ rIrlLf >xυrT'p@)c w@p+DpЊ ||=%3ȶ:4\ve!$by>AYO Syq2xJ`l{J%\G'`Ŝ:{Ԡ9S8a{'dP+){؏l]m'ZNԡIA~Rzr9jAn<. g|J0A7Y{ 'npe_?-Gᅳ\ XNrYҽpA8WqIoWӃ U?.=Q7znFظ~ik.|w:;x]%F .x}hxxԮ,+(BFx:kt3*f+8'H6q[AoXBwRe%@p1AM X~Onٔmm˺fu3 %7y& 0 2kdtQ6sAHzC[{W5.y´DM*5/q8 Wɠ9$ 6eJ;Evv;|1x"r;ٮ|oӓML2)I&@8#S%O2# e"Be5価[6lJTMǜhoUriش[=Y{v{vcRY*7rRD(F$!hRs)lGt萐H54o#V!j*V =0|;`&:mL4ZF)BbJL9ByE@@]S-`M(0<ߖc߬KkWFM>.ow7ߦL"Zu|& x} U`"&A9)"~I"&FT6z7dOLgG#ϓI,CA%cK/Zt`(Aed2`6W.sfR +hvZW^=c j:emf\΍3QHs~hjecAHBz/&aKBebV(8dehQ_~380WYH_*YAWoz>xJ()P98hbdSp$wMh4&&)uVqJIjD_w *.Neg7b ?GaFT9`=@ٜ\)a6hK1\[zG4Cn| t6x0i4/;|)dſoӂ#{#9j8Cp9g8k8MNNrу^^YC|ݺ \n6OY!la~xOWͯ_7Ơ~wp<`W'et:[ o~z?h=myASǮZx^ tyKLv_*SfIQ.5:~_Y HcsV%]Y"TňIF&IQ}gZaZhjP/g`]y>g9!OH4t4eT@ҚF/6 JEq.%eKuI.6 HJ Sk y?&<6ƋG.!iK7(Urk}r6~ hTGW` \!>L-WJ){zp PVNgQ[޶yڍ|Y8B-[vH`2u&s=xސ2:q5~>{q'tOxo(@%JD5ҽj `L]J3*#)iI~Ń]}0Nԅ‰9;ġP|¦> 84<9O]6\:S'*1F-iJ X%8RTۙ2}\*kߋc=ǖ; ͂eeζq#_0 v(non̾Hf6h͵.HgfWE=H6)&dL2wWn/pyjc'2eY9BO˥;=rgY(mNby gUl]Ǹ˳ Ol)Ii&]]"#*N,be.P\2r6OT5b`-IQQبv>!a-A`Mʵ4\ZFDq*pC\IiHP 6<Ppj:P%z+&!EW(X \C5B !چ=62d* P|2*-z+%4 \`ʕLZͻ+Ti> W(؄ʵ6\ZJUq*9pbp+^ ,L˸'WzjuK`=c 5=BPp6Ւ+ Wڮ Uv-p,bx6 W 0 P &֨iCpC\qCTP`pr- &H Tg0\X@$( Ԛ'Ru%Wς+i& \`y0B6\Ztq*0wG\)cUJYB W `AP+uq*fz+m!]`-™jFPpj]Ǖ!#V4we!l)\\ACUBz12MoFAD븪'X=wUOnSꩴ{gpu)ڪpM@B+Tkeq*miuUpň搜`/5+ˈ Wh-r+ʛqtqc 猋z+-7$uuc*og9a/PveYQzkl@tԲA!TZ ]vT8%Rl``Zֶ/r*[eBZ5,B JORaS?Ac'oD#`#*gG ;:Vб!GKpru1F=AaP?-M]eUi\=%).XpNM+Ews7[coO'!5FR zMdiX1uSKzQȁ{fS|TFg̓f{*۷RP/[^h2~mB/ /?è !\*Z6ogZ|xYb^hsbŸ4'اF2|{D|rip#f3r e FPɇ+{9!Ζ< \`Cd0BT+T::8rUq%*vWBiH0Ppj|t*%WRfl@\`pru0BVvWRvp\OWP ʄ+TΫ5UJ=ઇZZ 6lerm0sWZ+BIq*0G\- W(X`pr5Wt\;$D*Ҁpe8OQn8EZ#+Ti'/WuOk 6d\AU=Ҷz*uǒAZWtձFiĚ`pr) WVtWtpԀkW J P]4rUq-{Fq-'%(Ppj:P>JX W  P.ա U =ઇֈg]t*@]\iBժBC2K\iB)x[8"\ZŻ+P \Wh NP. fնs=rXG\YJ  ppr- &DOQeűWU4=*i`zrU-SOeiNpu)} KƂUچ+Pk$:(5DW=c z2XKi;dʮ%W ]>_U<^iWSM5cTcM^džsM--w欥:5`nad>_99C<_& ƯWn @C->s=o.Rk?v%7^r>u(6^$.>}x0t|w[ۻ?a7w' S穔ibhZ1yj"&݉_mv~>ŷ4Y pǙ9FH ?cR]0 !ϔ>$w9ϴ傩2;=50d8{WL-A@5b~+pG5Q`-![/A;?Na`󣿬˗ e7~t]9F]TRSҍ[OfpQpQ ť|tKȄZ6ʌ(S#Jwŭ*j<7Gv@e>#D'̥D0e,)HT2hZYocK0'"Vxm{>13_./rS@^е5yU8|@V-β)P\TE;T|yxx -MdE4^,|vmbM&}l[TpMlE 9pϣXm8l~RBA| .K^X-9_ήWRĂVe矣/V|B縈>IhG2cM5$*jXtsߓˡ}[3 F/XdIYCge+BYl21ɬPlȨS]˨NFu~dTkD2A5w4͒\'Hβ43r>KHEiB.lNPpP c˹dro{삱8?wf%GSrl I%'$\֣.>.H؄{6WN^R坈+S2' "KUj;T\M\Hsp1D+l6/RtwQqŰq藃-rjP-4PUn}>zFx%Ƨ\Z)CT3-d"ЧJKZy/Ez;z_0[=t߮o/}ŵtv= '>n}hpƄq(`~kۼ6}P=aA_EzA/*Okb{Q8 Iw~Ua:*ԇ!{9/|7^?{Ƒ_v`\ =[CGĘ"%[Y~3CH%9]SUz8?XPk0Z{sא E =ጥ`*k^w=ԣw2M21-`<&7)('h"WYNhC03_}7<}]Wl̻Ҳ8hZg  2A#t Is$= -{8&Pox' ܯ/ۏ]g:<Vh~!rr6M.Kcs0%\^EQh㳡j $n>^g5P r`M&OIQc$b@KeQ`)U: gf85c2(PR mQFf ^GIZ统]{`C=U+:,Kv9o%^$̓ċd #ve_3AJ//gF׌osayjヘ$$0DB}^GsޒY" H@`1@+8ܑѫQ1-(bL!V-EE*(4$i3 ,QkΤ3%sx=b,9A]`-[RDZ3ZSkxUg@] +Koo7{C%|87:%W{e>"{&&^!!&t &&NP;mExc8hN$AK>' (v1ΕZo۾/&{ p]`\P`D-V^G|gI/DzlglcĪY;l'؎QCԩDgu/0BrA9]4 lҷ{ťs"IRz4b5AUI !#zj1!˴҄QklFZ`2# !};(3/6`|vJQz9qs-G;HG*ޠ϶!g²Ho=<)$qc( Dh,@\ ڨ<QJOu]2K;,kd2K6/ͶCZ4RɴA!GچBIjmi'ATB" | L #Awh;ESt>b"2j3S"EtSd'H޲ WrJA@=Eydr6nyljܦfZ{hz*\KN[n؃)~9O|ڸ9vБfGp[57ڦsp[7MmI(J:x{n5394U˝WV"ŒzO2vc2h\(GuhR<&p yMw H)xPLlWk'AO.m#ɳ v*@ۙ0dXAe!|e_[jM C!( @$%֤ѷ֤7.sWuzx1꣊z)hњ X4-I3\HpTź MA0TQA&*voHäPHIqqs.AE`j=Z74EKhn{[li;qtzkYLRZY1!KEdJr"xpB!0ёhäܮ/% -ۖ?\W=4\ve)Lz/UMH0%? ]eq(/\g/LgmJ}J("e(dyO%s2{\GQ@vOw]S8<s?>*'Ϗ``8rM*:|Mx 'M4I#N2Oq{7/g]J4+f|tBJ!:5eI6po??g6/q8-'7P?ݫg7#Z~|.5gQB=w_4蟞+9ц?]N*37-o!kfҌI3F kN3*Q # G̫wpp=]fydOrVFuf]ZQqIe>`%Fo}YJMhoT@,iꦬ) y^_!;?~yolʠZ z?]~};ːˑ\OU?fDAWV؋ӯ (u~ -*LF#Xb7/>mڔm;mE,mbYLpC u'B##CfA`-N3ü7Fx8HIg(=aJ|8$T;v6{ *hNq~d3kSf8d119JVV8?$֤2I+a]kdv[p`hQڶ~3807^jCgnݾ^(Q gCV"*mBB-ɝjIJUR0ѩoWd{'lE9?6,EoQc92!o'(:˟C\ѣ_FH䎍qg(rCHqթC\,j|^m>7|S=&BrcoRB =nϏp=r\g cU~K0*z /~x9]nO4X;|)W'E]hieO\NO*+'f0pw+߶[|a9 )T|mbak"Bfw}i[qo_׼bPhRBEkj7+u Y9DMdm+7 \RBȩEMMejMtx;dyſiHt@Kof.'pWԌ).q+w-h(٘ӳXlxpqذp1{)\W6Xm_7LR NC,P3j3{GqpLNx;f|Plbٔ;&̈́-wu6<fWOMrӝ3O_߽ԃ㲍APF8Bb$\AƵv nsMM0*%ǥyy`BEtLQ P4 2|:u'ࣃ1U7=#ғ'>qfqFT14Av^UyUC޼?ʏ zxm6Cwȳen fP-Ua5{GշT^}/p:EM.xX'^}vYwUqr فrczt9y.տwBsղpf7)qY]`$8d,,uv:#2"}(F򎴴o$'Aư#H"Z|BbԠPwsT2HќΦU#Y8%3xC- $}NH]rgW3ղ9[̹Seҩڽlkq62jsVIsO"yq+"ɥ<]A'&nD(8CnUPAmYCJ@d24@BAkMLhYP[#gi6wEq8W@Þyه Arґx~/%YvTjYmQQ$pqo!r/޿ue~uBm]_rhB zs މmݝvkWCϟ^KQֽu%8)7ZU*ϮȬ[Bp눕rZXQӬSPgאgZMǐz>` eV#2s#: yƳ:,G,gKN\>ƟՇ\볫_pR8;J dFH!Dw56 bfФa*RXt {Zi&dIZҀsI>%]2ʜÈ]ߵ{2xZoGQ,W'ZbrW= dqr#By1@!"bV\Lm܃[gjc>qW3(iӨw91xu{%\%EY|#.>ZLh)S$N&W&Iw= rs=qKKe\< 𙍯(/}ˑeW3{b9 fE~<\}Q\_^!sK\Ou:v$ 'JI9vP}h3cwt|(}v̽/wmG;.ït|&Kv,Z z?w4.wf56n&64OmDh䪟[NJ"WwJ?xlo'7\㺙gw7o>m 𻟦SfiCU*?+QD,{s xE폳9=N @ݟv{J*=RoDeJƯ,'%8溓|ɑ'G W2> sRɛ4{7Y^􃑏o3;(.G^3}^ߖSW^XLi7GOe>6W+Gw7~^ۏnJ{^/\QSOaq5ooq+>z"]|Ur;jo33/8 {Iq.Q^B:a%ޏwI:j%S*"kppF$%!LԤ Unq3$a g2BqxʑP/ksќ2\^ޗD~Y\\A/~;gSS>$GO>sJյ\ _ js9[CCcGu8s$ݑNBcUݾY}m~CKCLW@3ϣv&5 ꪮҮ5P] QmaƓs?čç콵D=JԣD}w3plFQ*U8 ,Tkeĝ+V^p;{3Y{e |6,+\PZy$MJ́\og]Vsw M^R{_hisrXد^W ᫔էW`|7K0Uy-}`mvVc@d $)jz-pk/| ]4׿v5eyp5 W? \)H|Epv5͍)ܴ6W2t"eWWepsaWWW#\pO^=ǃd% noy͆{|b Sv~,حǽ[@;qt6W%\iWV^{ůxrvru}~y5;E1 ܫ4&q ]\;~oV̪//uQHo{?΍Viޔ+Cqnぬ InPV>>0gg FL nۆjDZ*2ʭߨXp>zaD->/,Huf9A 3aGO#eon8c`/ONߵ[sDśsُزt5nIZR=qj29-d M׳_[Mj⵳k&!Ř]iTjmdiH{%N>yl۝xZuI%vE*loj˗Z%iKh} &zҝ9#&Z w/?b@0cQfp,f W1ZuFs]Z8L8,YG}!G-9WrnnLx#qX 1Q!_Ta")|ԎFT.m MHER3`ON|N9 /p``$d噠\֯/Ίib,0tX `+%s-!^ bg')cOM*:$R+VH!xH%V|lY݈ 77M/ph|֕xhه-9=|Y,OHcMȟ}.0W͐R+uP )%H5ƒ/M* zlBNbm $edy2z^XŵTP K%k N-wRYSM^bi:P ґB !dGo/C:K~ og LO$ˊOhZe2V)+T[ =EF;  qG.rfH LTX|5J|d,!mc+pmp-c;ʺ VtxXZ]Ѣ!cq_bִ؁qnV 6jpWzb81X&g-M4Qq54ֳ@*es̥:pf?YJg3h6V3KHan. [ F붱 1>9ӓ2~q*xK@3G/ mr0)0׿iJY::t k1kEaxQ8Fh31Y_oFȀofPa,DtA,g;2XЁWR YG$`#zG P}kaqTh '_؊BH"K܆Np=n 9A|k@_WՁ C :W(,2h #Yj5H6Y/E`Jc\sMj*Ycv]Bcsr 9EfCV'Lm@hY wڨ`~"ܓ/B :PfZy~;(mf5S\x -G2FJ6ցP4Zsr, -BqQ 􀕅源ۭVpJ7a"M ~\QacSP qϖV1ٽ. D 5 $P~F$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I Mih=IO٤JHWGߟRSqvAOu8\ެoϮ l\"\R% \z9%&!uU':h!,2- ]]yվEowY]7btŴO-uŔTW+0t!,EWLcbJ̪g$IW7)Eȵ#ǥTW+1+M]1ZFNpó&[2*unx+HW ]1n3δXQu. (EWDLbJgUW+` ;btEފjg箈j[&:@Wcc8!-L7ąE1.4wUD|e @Wڷ- t1f1"ԑuŔUW+ԕ+btŸrzWLv]eM`V瓫W'\le{6t_u|ñ3g1}Q7-ǫy9J/~gקT^߅ߏW'e37AW Ї.7?7TBށt*?<Џq zcw]gokxu9mFp[_h;ا|͟<2]΅m70th&0A3U<;PrjJox:|sv9.CM kr6? ?ǟjt6C5ER}5] >g׆C lT BޣqJ9L#haj7tC*t% \|X)$ZOҔ?JL8{5bl_kwiك4f`Lb؄ Kb3W8gtsQ&HӺ\r6Zo ]1p+F(aS]GW%zW A7E)bZt*l֨-V*ٸNWMv]3v&2rkv)֨+1a+^[+EWL;J+m| ,ṗ{WE//Å7=, E(SeYPڷ-`|߂UW@7]1m)sR]PW.8 ]1p1I>jK' btŸK]WLwF]AH'AbbtEX)bZkSzZ~ wE ]1nӻ"ljgJUW+UQR*F+V\A NkUv^|G"]1-T?dʨSkU9b` vz]WL ڻZ0י銁? ֚P҂*>ISe2X K2XFk* tUWG+X&Zgpvp6ueS6wz]E_́sof;5lQT!(1+b6/f6R *ԦO,)tlsNI(GTl͂@p1RL$J:#!FI?`YP6uŔ%1UWϢ+e銁1IRtŴz]1W]QWCuɈf1CS|j=qtQNq+}bJ*aAx$fnbڙU)}P]PW銀qbtŸNv]1eUz`ѧcc,uU_Pޖ6etUF+;#*-zk|B'HW_PBeԮ-%V+g HWlS+uiz]1euzW|ܦlVWw_}}`Y=P4}r?A0uގ wS䂵-Fo页);(_B=?g׿=ݕ=6Gg !_N޽ꢽ|_3MogW\Roˇ?vޗ6q0&w}?G$澊K 0űд_}m^|& itA"Z~E54z|;%X볤=C+b:DZCSFj K1r~!\btEv]eQuB]E3xA!1" 6MW5*9- '+E'EWDlbJkh֨xI+ FWA61jB0IZZqJZS})]]}>6Vket\ wi]Ѧ(seʪ}zI{1b\pRtŴ!ծ+IuB]9χC+ IvjemqU_)e6ky՜Ux ʩë=ۋ:LtTV1j!W}FbQq: 9Vw  7E)ClES(Xbp ctEɀ]1)bZ_jnsI8'FWʴ(lezt!銀5btE)bD3VAqm+կfcʠ֨sLjbtEɋYFنuŔI֨+(HW+1SLcbʬsW/GWcR ]/EfP戴V]jߢdtي+]j!CeE,FWQ6CbJ ' (FWE`v]1E u A"P]1m~)SV]PW! ]09b,f*d]WLuj Իb`AAE1sWDm]WLjJٙ Vg9+}=OWLk 2K*gh銀 B+}9S& QWSI8[9Ab殘vL=c9|FWeޑYTWqe<\W7U]S.cd- 9(+N=@ 4e4׬àj0(bN)!6(f͸9Hb;T?#Ȕ{ClEAMj.|P-g|bJ* /E!n+MYXH]1pHF[2ij>AvtEaee!IӦ{WLYBճ*y_|+1sWL[BLIuB]epJ9bܥ#e_5 -eRtŴ_2Tѕ{`ѻcڋ-ݻ* 몌v`el+ЕS][0y/HW @7{)bZ^WDjwlZ|W {yu1l;k#3D.\3zuP(eWY(%M$ݴQJ6i9[8e9ES8q .m;C{5޾Ʌ8w947> =۫[=۷ois;o\ ΂@ cǹ#Dǯ ;߿nwn!ڷn/xx g _\ן/s<=fV?o]kG{}sFܞ6O!~]QۯN}G~~}qv}'))c[{]8hbfRH}OA'ۿl^߫ כǦFօƤ.9!řHGܡ1\r# A@0]+{ݽ⮺jpA6}#}O݉M~!9~?7Zmot9rwi]4 S{1k_ٟW{Njv C{*_7~}?j,Eb\.=f~{u˩Ӯ֣uS'8ixof?m.v|l]?P&7Ƌ'g랯NrDg2*.7g|Pû?hF$;@s: &lt}FMvm#I+Dp0R^l8bHCIVW=|8"M5%JуTUUuM#b:m+]]WcQb"'fʷ 5#: ek+ z nћro=)\%]z,XFǢ$d_퐅C~AЫ:/ N}",R5EY1R}Ʀ'X.-N6y ̔]9̔C.u}a|ܑ+v;?̽ى Aə4d` _Ί. npzL`YZC4rR۝޲=v=dj;_.(5 I!*R@E锨G#,Jrk)6׆VzZġ Apݜð7ry95!BNzտInlͰxJ{`];ڽqEܶa{-2XI+%_/ˏނ)֗ד}mf~^<Tq"҂VwPy',,.\8GM d4>]|Ɨg1cBpz=ɇcB_Wn贊6_э=b~9!e?p e)ʃ2FZ I/{C-L%ʬbN0p5@sb G5.Tїcٯr^׳Nv3 4C+]Xl۾rZCä^0fa|BOw˻s K3RχHܾGA2jݡ$J0LOIQQb$b棐e2yQ`)U: gf85c2(PR (ugoY@ǣ$^~ s2`_yY,=g"W j2T;U*E+k Qݡb;>ԺW_C*+t&Y=eFD(C۠3*UHc$B:?~`cS-JC6<,S2ѝgY6W*,T-:z޳U@AxȒ zcI&΍6hpUS" xTb2y_/Ն;뙜.Δ<skluvvIs*֪ -S %z1`XbPG-c!TG`ZHTK 2jb!gBZq0kPIEIPIiJ-$̾ `\ L #A;D+䜢G<9(y.Jh̗6)Xf"'H޲spJA@uј`3wg{Lj vжpVQQNk2>O@tD(BߊVO[=:K]~돆CkЁhK^y ժ* Vxq-JB pE-p!zn5{Q/ySe^B5s &eJ,>PF,7J5%"AEo3o$/\\O)gn v.0k*dKwHwa*t^nGӻ{y7%jn@(ӧ?K\jpM+S pJQ´ew ?sx<"-.Ô\ lj1z81*hPlAeelxj 1 Ds Qk$ jO8KNPXxl{hш«[*qvxz q; {/ڡ?AqLGDB 0>$d:# {Ew~_ɖl4MH%r-;cT[\+(v?܀o;vӿd+7yq%Cy^Z:1;*+D\zl3žīYm Fv@QCԩDguDIH!.XR 6RRRƹt[$)}&$J")!d`$PO1&gZiSi`T;ہѺ:HCHCzpUy+,m?g"Ag>9USys֧ I8I+Hcs"ID ɗ${(GGBvy!5/uSE`{"z.QzD `2* iMĠ$pߡ"(-tb_I,hu7Sov垀b=|)a?>V8E=nPa>,:ɀsDWBhSy"Z D:EIψ:.ԉ=ǂ:jΌa9$"FF B(+`A$$KMQԩb.>罥;FwJuQgHdS4RjoVJS> H^1 HVqP Ik3D>,XB H6Lϒ"8hQI4k}/wO~7b}ƷGu=j19'*ytG&Q͔dӺ@4 K҇kv %󗲋R*oEZ4P$)zh;粰{P©Q9Ih^QbLޓ|2ehz `F=,:&%4DE,*KhJ[bnabde ۶llAmGyl[HӇ oemfY+6O@͠_O^f8,4hPP1B9D.ڂ FTO*GJ ;$p"#{Y6\8:oGMLXVmJouvGøb.&Ga'tV[A3$82H Z (JKF蠉M8!Q%IFi,jTo$!#2DD!phk"0j%(yK$jЮbź6N XbEd-",bgexbc tg]#*FBT))+DH* E@&D`(PTs0!'̈́tE,֝ΊjvfuLbR]"bgzPk:PC7h#ˌ\O|ga},l~&lRfs]m=f |/ݻS;:3jd5ٍ_ ݕyC%(ΘV @$%r_] TץK(۸38)Af"JL بBDx}>p)&*ȼilX$_S̥aRk(c$b8q{cϹIqQ[%4/֝ٚ 9uCKksTjn?ގzۛX !]`嵬.wQ..KT\D$''hbS%’m:_u : M\)mn.0T82X)ըP@q5s_@~!L6WÍGA+ʺS$>*̢,hA@Q$Mu0 Ap:O \NEa]Ij:Ji!$$7%HC!xf@2 ai0"tTV f 7O*|A(c`(']-[6♔;/_Ϛt#? s1 @Mў44\'F7MoH0%O(ws.աt,Gao̟4ѧ=J(=Nćޚ;#`u.NAw 8a'VSN\]Åܞ4'0Q\>1s]*:| W[~8c#6iΥ>P~bpN?m5d6K8[trqrWZ%:5; ꓰK!lp^.|11~kyf׌?AFn[ѬtoZ<&\ɏ̓7Wo(ԝ^/a?Ogۻ]  4?=w` %hkKBnn颭֌cYF Y+h8Ot}f?i2pA6V:A\Cr篣qU_hR/GO 퍦kV1@ e^_8׿?_߿ϟ~w?Rf>W L 䫮ۺ4vGu@Gt^睛愖jۛ7hh&|v%)9vϢơ(z?_}y3?ًf'o:ݭ3z{q'mTfU4zw9\ƄD5c AiK|ME⠧%S>y"T;>2Dad6RtQ6čq:QzUqLQ2|e׃JjM4޵#EOw(#};ػ/w E2cLv}nY%i?m5]MV}Ydd)Pn쥴kEMb;βEU 9,L,g/%k2xH6W=a6~:6;/;iz޽K p/K1¼Ֆ2;J[<>FtWjGy)e-(DxDh< C&ѕU2\2,eh$af%XiR`v ZbfO?,hgu؁o3+Wt <*Hֱ2El #*w9R6 1<(gzB9T&lђ*H0,!Yu*lqA O'Mm.1Q@p|[}RF#*$cS\lm(+I qɬ0{IFuj1b)C#C):;NDr"47ݾ.̗ ~_.rg@o6c+0ڪe!a Ȅ*+%ElHIȚl%dɽR[x+Fz eƀr"~3yWŬK{;O yPz>w1/[}֧p +rT|#YJ'ւ:EΆn~3oK˓ۮڠP]5=[on>IrC?iF.{N~w :tNˣX,~R o_:)Ź7[d֪|qd|2ח!42K=7'9HYi\dR1`#>跤O#l#"# WX%Y7 5 nO^HiyǪ/Í@팾=tL;m;j}=& -SStMykQ9kS\9FB!=X{"{m`l8kѓ!P xH 4(h]2'"V8[U!𤒓:e4HR,&M'&fN J /K!>manvǭ/ȶrh}Xn~F+yo گo[Y+=->5}ym{Svr>/[{eh=t~qZ;.ʆqk_bލrj^tߣ"}}93_UA.r; LxS˶glMQaHչN PCBwCMQb T!2d#]R>$V**5ys|G\wV_;:OQIdk`+IRJ*9lB*X(SUfkmɶk82~`u_o(nR=D[Cޱ>O /Mn3AʵT5BTfpȬIVB+\#D.:BD8EY$iwJ*fHxDrd XLJ؜ :2Udv2q9ح 4*e}nԢٸ}%u~g֦'OX\}"{tD_}Rqr\ j1lg'u2@JNi^UQsEGo͜eeP9.2;#Rqc^wِe2R3d'K|'c?{P r"!IvڧisbN3h(J!Rt#r>*#6Ievhg#lGy3]6:9|C~lȦ&]s|V]YGsqCCUޛqbFeW aMwў:JjJJTV9yE$_ZgctU^^6;;k@N^u~8_M#>-벢+_>K`` '|=a>OZW,)X,80좓̵#tu8Wtd۞Lz=|Q|| H7Rk-T:oۻ\sWzqmy8:lXv^og?uqo+?\;>]=`5P2sqy~-0?u}fu^q >5||>|CC~>lZڝd׷y ^ 'Н\W 7A PƢM~î&72\^zo`>\Ϧ rv;bq&9cF^VFğׁ+1OZEgŏLzͲW׿wZxܳmfwxDWjPa;~@Tô>&ȝ]uz9IYݟSϠXHPʆZ :Y%R4NTdN%iK!aVw-o ݐv5uG;w8J\CNU,DhJCnEd>hgKVH02S燜|Thc](|zJ)+έ4E+'M"|[ c-xI+ݝ|i|wJkbB㦔v}wJ :Fmȼvd1]Z VH)ɪ Lo-Pji;]+CP0f cUIh38f"F'VFgaâX9 wrEءlQ~RWCuLJra׾mc2C(R2X>CZD@rWՓ, ԦtA 3D0b恠+/g'+Tl5޴`<%c"BJ&+?}Px*qsT٬%UT&R8' \ mQu :Q&RBr3D}d S!Gi+z+QJI*P(2í+8}B_+ׄ,hLڈ* !$Z4А&AlR=H@"Qa[fzRȚY# aF*\JUY X.b&k0QaKԉ Y\wv*!x$+v# p3Uck\4@)5>0 @PHI^c[X6 T0$(*X?A"b;7 ""ʉ0;ȃ0O3)z\cm8l"H@k߀LҒ!ڳН`G4D3 " Y"F,Ko${Bg_p7$[ 2r@ y7/!"F)@hV8ppU>6]jl& >wD&%+ ]v q18- xi5:@8/Seu@x'Q,(B P RR]9_hFOpb!n<Aq %VSnkIμxal(*׳e?T6l{D[g?b@0}:L O $?z&H)2# @B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L.&1O +t@ O ")B&sd@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 LZ0l @  @χ $(@dL d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&гe j9Jg}2L"O  Tdx&гdq@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 LUZ:,m_Դu[^֋/{dVw pL[L""K aq4OR0M@ҙ`QդoHw",WVwNy ụvͻsޓe΀$^LEJ4H~?)Xҩ[Y~\bDn3O5Ju%*ϣؖ.u.:Cx LO)͘rq@,)=QGGЍR7;+fz|ٛ+[7j֜QK[ҥʔ(Θ0P6%<;E ,NiRk)g[)|㭔Y{G}9h6 8Z x WTmd(+d8~ɪ!@/PF 0\J*v^>MϖϹQ|>tOxn6`ICeR"7>?vrZES`vaG^(]PH.ňnjD \p3WN? @ ~Wcwa`ھW, dBR_->t>6OU8DIpPF@xKcm@݀C?x Cm{¨am{T7]gaE?ۧR;Ǜw, |~Ec[t&˗CFp+ y̖/\|h^thv + 9ץ/!P\LD˔1[>PFʚk"˥Z_𠢷XC['GA3+YB~^Qa U2uB鲩]bwnb/F bH!)Kӭ6Ɵ%CiDrFQ4+Y1w ҂cfWR+8ܑ y^N*[ &EY:TVn6q<W?lp)8SgtaBqo>hD>au[_[|Y'tL6Ӥ3N*w"k2icj~)h6PFb|8"T(`_VN7~IF BTt"X0cx#w5$߀Xjj $loV?qQCԩL]2 60ﳢrRhc9,F*J jTFĹݴBG@avx3ͼ+2 ٓ[O?Rv'瑠T{>aU9ހ1xhYQg7su17%t9&F%y{F9 aC:^FEz V>wO-0LYmr~7,on[m x[S=Yp_qWb!yxm~:_V jp}WV3Xc0]//h6nw@5Geli=bjv_u_3/M' (ˀY69 D2i|!`cR^]W50H+K{ iΌa1d#$97F *DEܕ,HV*KY) ˞)TLUgςޝ&r#^WlFB0}uDe*%i$ $ɲ' š,1[ j%5Rixe ggT-0}=و~ɶNTr+;h܎xr}t|}1wbZo ?= ;T7FjwTFܾTZ˫b Z4&Raв sE}R©arng74 ۫A2[lO%YO1K Ff*F(7@[YX)PddHI堓tC \d{8+f+q"jfc36 r0hS8w[l?ƵXv58 }x Z`gIq* m2Ԗs@Oe#tfA0%ʕЍXf H* Bd[ %fY 6 >pd)'asa5qmPv/X,b58"EDHypes(.}޵p@E2hU*rbLU4Ѫ-b5qL%jvq'1:iɡvQT"]O e[νaq,PCD"A{a#U?rK^xȵTƲ~S#(6*%]xo%0BqIkISx,ƌS?ʑ6&R)C9LO1ÅLq4M ,D IK2F"4 '76@iQ,*9ZkjYR+_ip)8tx|7nbs7Mh$òs%8# 5}Р,\ژA"2U@K'D th|9h ²NQ$s}x}|l }um.'Fϙ]IGuLʍbL:/{eFpD\9TGQ0?kg"IBmI)| lf̗ QH!3NRYEN('R݉tPb4\LܹP AQU KJh!Gܻ߱ͤ|=6IF~ ;2d'FȽ Β w<}FWr=N9ؒ<]yBNj};vaqͺdQB68o{3nvgR2? 8! ; '!7s?>L m"@ȉ:wԁpnk~x7?]ѣ4l*T]Ώ#ПsajsTޅkMɼgYo`/Ng=5BQ S`r \<.پ8q T߭T7Bm^i_fWo/?xn R+=w0m?,=k;8aWJb](-+ Qqb_arYGR}wS*+"ؚ0еVmY^ب#mbϊĽ|q<τjGG &B9uʳfyo,s#` 58#CJw `!ejWƉ@RYJj=xO/^̟UT}P65$; uhgtٻ6lW? v-c$L ̝yZ- ɐlg0j6)ji"K%d],ߩ: U9BOD3J?p]pUJǔ1qv7gZFX|T2xZn[ޮWm}l\B#IB)U"-Cb)昉+ <'~A"x\E"wJqF#8!@16p}5?cbYC mԽ@Qmo 5,AVK5xa.A3*QLrޠKq06dJxɓcʑd"1 ŹC-9g7V /qȪOiUJ󟡷n^[+)19hxse%U>O|Ժݽ咿iBܳX _@4Bq-"kH$AOߍV͎ӢML]*.i"ުtrv}uyV-88R{ȍ߷~w-N~kEwayA'6ZO4o[H7W(t~+<-'i28kq[U4u/Yrco;q]Bm/6_Y\ܪקp?q>m{3=s %5@h>_{s$jY}iw OVcq8۞O"(2HzdtG%gY~=~MT6L#61ހ˦Afb!+GRaϱԏF%w0z8A%G HFM.8l [L*Ņ.E& \g@!  66$tsMB;kMɽ؀lY1@'Vu@#fǃcDn$:tL#- p&@:#`l(Rem FR^1dR:d-D ǒ"K(" q]Y=PL.4Ƴ"sVP&wsPgdg BPJe7QVZ.q^@hWQut-Ӄ4+'^D 9nJ2QL5Z6,kACLDL#5QS  4x2XXtL k}1yd 4FAA𑫀D({!Ut:=ځӨKNet[m8'j 0MWeBZύ k7<MzԒ~7s?;#ei7'!_TWww@V hgӳ\̒!x"$ MuBn(p7$eF}RoEƿD: hH9i- |01x.I렜5Q![Qynb[ zig^7w_'*z-OIq$GԮ(’a!ht۴ȭT}[؛?'kN"`]A۫#n0DM曒5kWʈPO.[ [lTjC..tTOiYRvqF.iICWK<&C g f&'_I, 'ߑ+KJ &{1cQ۔{G'JZQ|BS摰XzgU)ie[)E\ň&s#w; 6^'ՆsB6ǔK{*?/rٗ&L_'zY͗s1~nc&Z@Nn»w=}vL\ҭyNn r5蚩+@h{4G;Cà6YfC)E늹pvKһl]n7-ﭞyfzM&[>ݞ;h~2λ_PKMkn1kmܯHs9e͝_Oswɀ[Fpt(F6OdKďC^V9$c>&|o>N9܁=FMA2Z[y#D<3Wek%!+u9^mHHNzϋo fy\{~,0ݲDPf qjgQ3w>yg̊?[[ isԯg 鉒 M#qpikq"%.sBs`rF 6:Y)&Ey-*61ĽȪoNf\޵BOlClɢN1g'c|B@סjAdQy0+,aيMS7j@4sKMOgqcPAKGLs9ps8q=WϞ- . )v`Y>eh ܐBЯP#7jp+nd94Fa&2Ȱ)7]D Glt;3ʈԑ]֨~4sగhta0lF'=1"i#`\"N8f6 vC؁L=y%]2wgc^vW'2 tՇ] cQE4R6slmXD@k 81XTwOEXT{밨l2&$DIx>I\.Ő!`eBoI{ƌ7Av%&7{Mo }C!HFrP\p"Cɂ5,_Ns-QzJawXy@; `/(UGܨIYE$# N 2)Y2(R!$ҥ"Kh>uQ@P1sq2ޒ,Pa))VԀ$AjqM 3 Dc*94B !'nsBhsdR@vwT6%cI22UΎ|ffﳂvBǏK@fRg[ x)'؂s IO3 Ȣu >@M|) ?$ֆsHK-h qDdYjO]Ig?@jP+ v`'˜\L&hiLya6$!IeUtCt8*㨮˪.[AxxAD DNFgpAD;skUH!{EL;]RS FϏȾo M,LI-QY<^dədGKPJ+ 2'G- ٗ;TeP_!ż(|QmQZ QO IMgUHՏ܍|^ap㒻"%7bZ8}MƩ/r<xivw'f28ш~)Ly kYee gPף||.U[RH([[k֥}F Q{S;enQ9etXl<_;z-\.f8_21ԈWތN\dj8~6Ey;Wrt!4n<(2(Ф7A)<$rwI"v4Iל>sDO>:"+`ip~k0Y#%nܔB. A;尳KawKlkbR{~.~xj]($XDe(|㽶sΗlnolbwּ!\'iG;ۅ]57wz~{}},Z+8|h.FS7x_ƓW 7/ ~٥H۵ Hoy!SWNC -%`~o^]K~wڕ"+j.`i3H/I&je*_yR_4Qc~{JշY1Xdh~럄M.jԵ|Qh R\{{}L:qXmK͙0x!GAu(R{YDB2UVhC H N2#+yc:Hfu9DI3*+G4lqY#Y䱼,BF e)'\f Ql!"H8(^D˞LK˦qFv읓qVF5o5s?`r&~+3 4cЁ s> wmGyI66`# AЗmEce}Ix)Q#`F<]OUWu®@ ٳi6l,)JwcYBMY.n[$yd$#i%7wErW=h]P]TjQӍV6Yha'H[묚'j'zf˝܅ :s[<!k8fW'ȘQ;%B!R*XV*lPQwژK>F&sm`ThU B"f!ƛ1q4]z_/&t hfM>}`x}O>s!W_x Wiy^Zydu!gJzsI3dɩMB ]:7+h7\;_k.ynz&s|'eC90.6J}hx)B@|C LUh;՝?)zƃASJ@B~H22{͢ NR.PrZE^h0ozx'rZ>FO!jqMgP:nΧ2?iءgBGoWO'á}G:8t1GKhSI,JNTuVQkXsG2b*sZ\쥜@X7&'gC=7=2 eN`%X) 3Y d)m5 g׷Uyo Kv:@ۙ0d”d9!`mђ %6iż ш&gEwkY#{TxYBJ2k`:dž@ȘKZX =gmiB`8 ь*;Ia&QV(cGBD EzT moP=ɱD"gONRlˍ>XnhyT 1T.eb$$d QF&&ϒ \Ll0jVL/:q r\ٶ9˻,uh(#KftL[usCTE̬[V < +-C㶵jm\h0uIX밨OA G$kLBm#Vg3ˆcA;i=w,_uT[>wJU> y2͘*B2$[*8 y tH^>>*T" -iZZ)=8Rul6Z=mhk6w8 ?n4u2#nŹU|p}lkSak$]SyAEW>ø:柴^HfK޷q:$iP"ըOjx~ƥTqK@q0AWL]FKe<1^xBj }6m5C[,'ɤn =ٟl9E?gl4{5='cR&V>] oƤrx3M=+Z墚xӋӏ{uAǟw_۟9Ϲo]珴%~ y( ?|ӊ``~SKkMMͧ&4ԟ7-*V7P|H~ Y[}߿DZḝjO=aIVq_!I/^]oT9%5rL ں{k7Ec%e'Qr76E2 $A ̡JvmV`K"Rs[HT&Nzj\5甜sa:VHF/WdmP|Hٕ}.,T4E6J:;שrsR+iE}΋or%g Bw*=CwcN߆̚Wk 1 ;ZEҐ  pN5"oJd.iU{pwIwsw\~~iOrm6B2tkf5YeU9Fs&ȅp!`" .B>Z3lvS:Jp09A?l/A;6UrKYLxvKINj*'gl=R )o 2+"E-R3Gp3F8F()N1 X%܉5Njҭw^;WWJm%RIR;x!闉뤾ԟ^) }+b?jv(PT۲ ^ċ=}ЌGr5rT7etAr4hd4f8NKYύЩǫC={~Bl>>^ ͺ=~f%p|[+l[Ԇ $׷\.CyWqk7~?#"~0+whvOtf HA柾6ۛMڻ㘦cf3X?>.rXXnhЏ D]hS}]ZD#!?ۧwp?hC ^<Ǩd{'fn}OZOFWW%`/,W!{|gJ]e*%7+Vn6lC\RU#pn$>Y7{s5Cl &5@l?7隨͚imHb|7m<ocyJn:%LQ2ڋ!|/!mM}1ڏ/B=r/_v#s[̗xi{=I1 m%.!z}g:Blo: w\sytWf`}=xd߄pҿafv7 ,}v\ƤwvłS6Ŕ-Mل%;u* }!'꩙|oo:LS*ZN5>MQ4W}pUi}= 43 M@f@$pE:?72}j-˜9aRuf#,Gt%PctJR)tUQAZea ;DB\}$&g29qw; BWFٳV}JL:uG8;>92OzZJmu*t u[LUAE`Ɖ]k"l%VJE!!,R0XYP[PuakCo8=j|rA;4jm#R8^=Y{Ñ_JZyw,LpPK)X/eK DJ1b HuJoCPf֔b!0xR?+ LרrhS²lD\dā" (WϚL7Fsg*d!2$1J4k 5I%FGVE2x_YV#g>ly oQǡQTֈӈF܊~t+iF3IUcwիăZKTL9d\ȃW184I1[ ܤraF˥3YJWOx_lb`T~4&4lawzߦEn}wu{_e Wh_WmWM~BͷM7ԱVAgTZkm#9@$H~0r{M"XoCGDX"R~CI=(H` {,q4Lok.4vAQ$, gXt1*ygiNV6&"Ɂ9㙱N[DB؟l̴̗v[2^o/7{Mʳ]63ӌP  h:/)N=*Fr|2 +U@=6U  6>^s')nV$FS|_͇%y< 27NMSJoQz?G>.gu/BiMfgoPp,2jpޤcJ(-G],:,_ E)V 0//4{1@glGtNW<wa`GwNf5vtؗ~Fw81ϪE]2"aoFѼv4o01;kxod3N mb:2}flZglMdi׭n808"}D>d23躶5guw'1% |2ɄX6Pp19JYd Ĩ^]잡iᡙP+*ԩvLcק]Ң{jvƌ9k$\bm5Zfn1!j!Ꞇ1s1 A̭G %(6D# =]p$VQDH\EEB<Yv!A!X)0ф麮Ĺ^[hҟ+^hIes2[죗uYd^0KBE!> ^JZ˹1n6)[uj_#(#=rb54-(7QƻH=#Uc|DjO޶7JlSk@See"z&xR*!jA.]4(W:b ;7unE9u_rvjqd :lmO׾3뻔08M><1_ƒK"y*Ki' $S ܫr6L*eL*ms }__vDZPwt=Q%yE򆩸QD jpR֠sӶfɁ}Du|z"*'J#hLĠDXyL!d0M`CΠV8>NZ~/y &R@MА ~“$e@.Vi R]?qt2qrYg:~QF0 (sH+V*@GHrv[VE9(-]Fd`]ыޛ*ǝȶ_e,LI! Ltw2 C&+Owh 0VS4d[P[0ܟA[ ؏ݐo6y&Vf R5yڒ;A$}'LQ*kkAWgJ=4 ˨WVoNjx`+e:Dq$ !t\AҤŵY#skXfƽj6G`L*rHm}n(׈YWKVP?akϩſ?Y鹁'BI$='II$;o@ 2/4Q[ &l`U C10t g^yW 5AqRP&t|u)sl\t6:)oT46uj)WJ9(+IѺ*m&!FwUL)1h9UA3FubN!5Kgs˽IsJà$u)Ѹ~߼c۲:ؤ:ՕRt;J:ygh# 1>1ڣ4$/<aw*Wp\9_,*)^l1 N}W8s4Lok2T]ȆY:HI/xRӜBםNϥZ5-ޤh1 h"`Q]. E` (l`I 0wbҹW^}xgG,YZEIsګkhrV1=K3.(ewKY [֖'VBYs^"@1ML 44!wKP9E6HP'.t sr,AGggÅlimDk^˕'Kr:_t3DC"K䛔$L!Є"YkR `2}6G_u`Q#n׊]h(##M6LYEgK,=j9}=6 U ngqJ늱nF;OA G$k j:S̎eMCNrnAZ?W61$Ÿ+Z0 Ed/QYSs1)$C27>) GcoH6yiVœ"'@p͜`hyLׂBN!~0Y(dήFO4QX 57šF*\, x_]& 8j[njc. >~(7s}o^ôfҤћFg'fih@FY2-Ir8 zdpOxS<g4 ♫ft+S=T%uï?;wo/}{~}{oߒr~q+$q;p鿾i%]5MMڠߠ]źvo6 .s&R Q%CrV0Y1芫 5:6g̣zFb $tLQPGM`K&W9Π  t* Eָ3z&U?>iB?8.^szţ@HD` .\C\gj!Fof+0ZSG P1~qy`}? d#‰W^uu9OwOs>;!KX%?~GszNr2msmU˥>LU_t3O|sLGwq/jZU[HA6゚~nCvo1d*v{}e]=h2=c|zm}ⷚ~Vy/=YҿY}P][E:2jѧڱWYR2$&T͈;zYeu~jMq˛u28W]!~칞Z\aw<#=n]Pg;%Kn=Xz؞:Y_"qgl{r_{w= 3,"` b_?tOԯfݪzf*XXnLkӉnH t'yomyC"t즗,SfɐEXqۦ~@uX5x3ֵwR짨a4|;pMX`?N- b`9wd6Kixqx9xŘĄRXR%j+DΚP0#)_nuYGRknFMo˾itp<îefdI *D2wiϯi~n";?F6i,[ [Ew!a(Mxlg?h;{!bcM]A_YGP Fm`%Hb%IŦVi-A)(ZG[)X N ;R-kP5m&qt{HCآ_vg?_fc?6Y{| f3!~^!PO4n\^^7@.Tؠ" -XD"Q篆:{a~K:KXG 1tHsQAFek%P*!k:-Q 2.:Js 1z,0JU:4F919c-~GP$0R2 00YTޕh7L :wv`"8mSC%C [@tqײ)@0XPZ9ǑZ9Z+\jl9nqQosŷ.ͯ {r2Y|y0f2I|7}:OkGsCFkaiꕏdJ{D-Z d8T_-*WLT٩bԥg4V"ӁIi[ ʮ  ;֞8=c;6ӌ|%E;_G_wi79I_',oL7Ѕtezy)"`hRLE[vBly@HQ( dЖtƸX9&5"NTA~ 5 TMf >±pzBoJ"u:Zsc,oCʹc^|bk*VE`L]0D$YCqh8PI0?MʶERtZ"GȌ $V8L?#Gxhy^3rDǐq3} N1*goR61H#-[wRcccҨ/8TcrH\JIZ8LSVM9\ՑrS8fZr_~Q~q)( `@ل^-N ERAM9,ȃV(~9Pa38.Žݯ5n8ũ4>$16~L=z7<4?g/w߳?g󭲍@ʾR푞=Om/fb?W2.!6pw(N0ڃv,SRTDH !v\PR.M$ e7R<]!YQ,èA "o_s}RrTp) m 1D6!.Z$CEfFii]cwZƙ,h䎶(`g@&yr__J^kvX.fկ۟KzQW6Y/n?.B;@twyrg? ґUz6s4@'I>&g;#8_҇@ɽ`MS%#\%z7Q=3MoW+ 7w2{3 wUq*2qb#F^Z?ڬa(w~zq`}~<*]fyNO<oxĩ yPU*]wjuQO;: 0"jtC>(>>i=f܂~3̷y8ZOz#֍-g&ߺ뺕qL57zxd߻IO%of LG9w P6 Y)@~o8P~CDDhхb ^\@iUEڤb*!D@@m DMMumѭBQz`o Pp$B(7Mk&ΉiRw}j*P אUZA׽ -V;?Z?z?kAYEY6@!eMq:+$I,SZ`F VͰWkz<4kTbB-Ǵ]ŮJcCEPI!.x!T9kT'QEZ/ec:c:(%b>%0A !9S,Ge""@8zI3l[]W$|N>IGeU"GǨsLev8ZsݮMo>]o<<~AK看 Ρ?2.# @!'k>՗(!:K^1nܶ::^#h#ISgb0BNbg/9ؤҡ6RφXΕ^H+pTpNnTcLU|PQB dֺ01%|@j(K@dw\Hsާ{GNMiص+k>gH0 "5 E')#ƜtE)g]!{Mq4&8vt&j!|Թ5_F}uA ]Y̓.^ s]6s@;JuhJ7٣w21)ZDp"d5"*G&jbY)bW*^! #%ZmSO)GʉNh{*()e*YBGr ,z)m1$ԚN8'JIyHiA]rUho/*#y>pdk歏ZcE6o?@[x_/wOTŻN#B9Ƞ1h; RLx-|6>t!& Fmfȷg]X)?):dGKI.Chɍo_^zkkaȠQR:堭1N2c|/KLlӥHEeP;J{<^{l3&_~cRŔ3ϞrzC颳~ltNXGBP;@ӊ9}uDJ2즻g~A/X,;ۯ?Cْ-۲ʒcy-/^p>X/ `J9%A`2) Mtq}Ye3ke'n]3Z>!-P=x_z%ӣt53%>A,d!PRRǶetHʢhmP "(s.+Ti ua`t Bsoށ3G8C;<>zvkħA+ݻ}v}A< '<ţ4[vy}N6l=VA~5_ >2T2vrWiW>1ḞCv8rN$0f'!,*Sg#\V9a*d D{vx5bRbq<]v%t{HkGHK;L71h떄iYilЄX9 =Dva&#hvDВ[LBg (c,$zDmU F$=Z NIIF'v}k0gy-(ؑ{fNQ9'pOqQű#+1_^b_&NO!àX (!2kvyƋ$OHme)Lң.u?p~ܦS6nڊ;;Ǵ1M4\i}a ( EIK^cN袔.=9X6Q{n&?+3N˺Zӵ#J3lθw:Ng;q#- -gtqθw:Ng;q tqwƝθw:Ng;q3tƝθ<\<+/՝>M}-Xy 5qp¸)!*lqu?;&ڳ%iUA9Gp恰ؒ>pXI9"̵Et* n6JPͺ0^a8=D/";&wic}ddY8߾ >|OwF] k䵲ؘY>A,d!PRRǶ1!)(u徠4A DQz\*4kP50j֝M+7Qnށ3Rޓ3>zv{%ܽv#_}h7Of/V$ chQBhsCB2)vT6:]U)a/l†>%C6Fa2{.$$h  T9b9M*vV, UTvOl(<5WF{L(m! &jk/B@=a<貴:;:i,B`v@n ,B&Q%GD0:Gv_P}‰99`ϑ|!T 6t 2\TQBx TJHuZ@dt'\t1•*bC7{[d%%T>ƨYw|ާh=G/?~O*&(pixotBPX}NG[QX›( k@tqֲ)@0XPZ9ǞZ9Z{mm՝G0{ K?$Lg39٤ T%ؾ+ldΔTzݿ2yj=&_\Pmꏟ&%粋*?W)ej-JhG2LD]uij%L9OPriz!ʮ  ֖YwJ3[L26mvw[U[QyEk /.Γfk}˲ wO CO>Gn!ZUDSrіk*[` -T)=sBcfTWT4v:Sy2#(fS '|};|ޔDE$Zou;L5s_v3 yN? UD =958"IPd/8Ԭ bu27ERtZ"{Ȍ  U($.n.=l&CLX}|f_ggt<_#>mL{U4Ip=qGُQ~\FU#|]~lDulIрg,(CdK *VITQ_~ bPW>p 5)0BǤJB`Ln" T~Yhke "sf0Z MR@$SNrEi P,e 9^kο|Vn,ۋsKO(XC`V%u`UĀy;+e@F>|:m/:{q.P,iR{Ć}5=("وϲfgm&TϾtJ0+CV&t߉ϾNRZHZHa2ƆH h%i'ZDpB(Pؗ*^aaV6 ,f!dR!/.D3r 5-l@֏N+dj'N{e _ "Oj9BSb[r2CPo0; Ffŕyԥyv{RHqW4Yxq?Ca%TF)a8 ;#(ԓ='إOmywwPf'd]2PrD})' }Ǩ^}anJ.ggOO^~uv|Br5AWgazMΪ׊U2}wrʈ.>i%dEӌwuvh95V֟esv+\z{&/YY{_{]rfWȋ?nϽӑ.1_okueqb3n {΅J>ݽ>*]~nP1OCIQ!ˉ#y)U% btg(鈠H,m eC]һ\,х\Ӹ껫\V?*5l`i3}Ibj&$1UQ QMww;jSymύA\;`UOMck_dwbJAщ-C$Y`jh:$Urm^Uǣh~9h_ٻ6rcWX~Jr<иʕSgw+gRuRdbL )M忟pHQHdcw{kuMWgDxnpVh>SFL̃"p 51N:6bXU@l,J(,]]f6 W"`epDmfD1 6D'RF ?'T"s<Mjj|,8(ZY.s6W'a'Z#-UzάR'Njnj Zt0qesuʟG=h~>xzgh?`ߋU_Dہ mKx_ﹹ85s#MA5=;A 3Cty!tnܪoQ֭;6϶ytϝ]GRdž彅t oc2s즎 7Z .vnY8y%v .Ⱦ -LiշۨEmYrBO GMK L)kLwB@E6gg}^^M#߀9]Tn :0ha3#(eà4(* X^WO-6-DϤuY\xW(Y5i9NLyRDy4*”Ս5RRҘdx"Gy 3* A8,s˼(:JǬI9[,g݁Iiՠ #[AG)hF㔹sr) F%R69kpI@FņS 8Ak4jVEzZr#FYG 8FȘhpCNX_OS_nZ,PZN lz1{82E6N]]\}TX)k 'n`Q" C} q436Ofdi 6~/fEkj6C02ɌH XO-{@s|gpiY^8atvr.hMa htd!9%꛼K&G=>0̸!s8XdwGW1^*=U7_+zoU֎]+w;ȍCY5W65w>?j4tĺ8˭y綻%ís}m WZn^dCyuÚGZuKsCr,w@v\_4ݕTsl׃{cJ㿨>Li{6*66( RÈ  0m(V^ܙpsًs,G؋S:91rŵMR)D-9AeQHPo2X@6&]xuN >ꮱzT \#Ab( TPRID< DBd+UC9+tdoCEEzElɪy)nʃ+N60tQ)V5$I2P]VL]G,fDεr2l2E@ku34RҴdp o<Q6ur؄C[iPFņ b_kdܫڴryAK ΦDœtT ]ٻ> ^ͥʟlg_ !9z-zz5J,:Bz1Jk k5ցpCm]fvՓݧVxJAeJ@C^2ѡ$mZ2/.{$с}2 Ӯ>6n-Sny|Vآ\&Y8n!$N!_oIn~A{w˸'42}]̏"=Wk|{ը,!|)oqrAAMk:bJ+zHeV-!hҖFu Q . wF}8mfz~ѷìw,yꂽ \|X`߂зfa7-,-Pw;}5XyPR֝{G b9]lι`8:L[CՏ߿zۼ~W;zieܿ~=: 3Cg8--&U8y9e9{ P"C>'rQ<ٻf++lΉH %ڬ z]ץk_́wS;?bk6̜O1M^taQ}]x\7*qDunP AQ7(Eݠu 4dSCjhReL6ɦ2T&dSl*MeL6ɦ2T&dSlNہ;m +M݋{q5M2T&dSl*MeL6ɦ2T&3]>ke.juQZl*MeL6ɦ2ٜ'^DDO_|\+wc.5h&HtV6ͥӔ7J a?=å#}wJ-h_Eq.sAeÑ]t v:)&茤Ds94EtnRqH"15XՁכq؄bkNB-!!;{맫6[vp!}ukc#C*t %Q4wl6C!*pXe 1R7$nRzTk'ͣ|-L2*HʍK\܋Jn1]E< rv60ӽT K17OG@3pqqI$.BFE)*g$0P\ȑj 9ٚ¤wkNt1*-X Y(g\zNb+hu7,T| YE2]Hcd&Z EؼDӐjO+U( mZE|~)S;'^D).0FfAR( L{bm 2~l9{7wKnV`+{"ǀivd{w_.sد# ډ\BѨ` rTNIr$5e$e^jUI1%x^G 9'6w8kzˢ̔Wn{v[7Cw}h"j|%˱|:LZT#CGKHE 4?I X+6AP$iIy =4GFC4R:᭶&H]t* d*yRpKZ8׽H}{kV*ОY>gɵhw^O_Jd&MF-kHJ@P*_yxr'+*M&pa*Dbp$Pk iڣh&+M+[Ճu7O><2^]3k%N>E攰ـ֬ᖸY Iq"SI7:WɎp$d烢V&jZK81K e)gz8_!ex_qn`;؇]AKĘ"eY1OpHQݲ(i9=5=U_UWWTTe.%eKuI HJ Sk )`?&<F jAHȌt;>xU#goQyMU_?v E>}4Qc q^8vq}+TG0>GJd+^ vv99^HUTeҥC@O*&2#-z1Dqc( Dh,@ ڨ<QJOu]txlBb`0ٹPb!BZ LWyKBPIy4M%ЖvHU hW¤04>y]EO^;A-&"F8c9ƘPDYR [s&8N)(h]{9L-"^<Miro!jpM>p\%ws=8Iz űCG?v: L46ttV/gUY[+eB@8>ZڽiNˢԻr~Uur9]ѣnkWdWb^-ep#@<+Y5s &eJ,>PF,7J5%"AEo3߬I9;8i0{Ɏ v&Ø0XDqfYFX_2B0h8cNzZ)@j=r%ߕ\[p)d. - siԚ:8(N 3x%LZ˛GTOS$ KsKaM\GYsbtrk|8;A#picB ȔDC `#x M!jrR/:~Qi,sm*0%OAhftL*H єjp(Bk%p'<6 {L*&.e2 :a$+3@GueF%K$4M=S#q$Wca|:nVKmIjsV`!+ǐRIEN(Mye.w.H\ AQEBF@`%bR LXmtu!) _ {A9favne_L{U:@]z s,~>Խ~v0x]J`l*yPry,(q<Ȟأ́! {ޑVS\]ÉST'{0U{m(N o%C >_rՂxma#N3bI$rըU+*.8ΎuKHX"8i+'%ݣ~E5tE yYgr/Ȏ7x&՛߿>|:BqHXmIxUh\^r9C#_WM-ƕ#,m ~Xq~a_|^iț-k&Z ^߂d4rm U?^J~IlTA X&BݺZT(/kmXNvL>s; u'B##CfA\'-N3ü7Fx&8HIg'}k\UKN0xa@#e˃JjM4Kio\$Gr>Ub2h,6܍ݾN}2\'Hmvw}E+xnU]ÄpiOj fͳ5l%jH xVVNZ].3-cу\8.1o=O嚤$ɡehSD;L,pFK>ieFH e"f"0P{ZYv)ccSrۭw?CG.V>[olXgSkb9%1ISQm؊,zˣ_t;?أưYo7lNggA]05awD~![փ D Bf >XEB 22k],qL6;d]鰠"ޢx4;^.ޛO-uÈ-!7x:]:qM Yw_gBP@zxX_M剖]||gs@۽?Q_Dn..jʄ_i"?='5E~aWHedWD] )RhŭNFQ O3^zčk>חGSM^~:]{ӝMknyÿ^5Ks?yykxz~_JYNs\NE?_}7'#oWGw/hiyw:z_}뢯%Z_ƹƽ16չG54[%oy7>({eW}Rv7ԶN;mˈu Ew\ f}lXor{9'O-`6g°_p~k676̳7tm턷1MLtwfwμSDpvwm-B//[R>uSE`| (~@D9QӔQIk@[.%eKuI6 HJ Sk yJJblFKTv}]^φx}`zVc: ALEQ=Wuٯqff<8=꣆4.pJm*2VD Ayg:P''jתH;U$?͏ĜUC'b*>+k š @Z ڲ\*dd8-i+cD3.1hyĄ‚Z͂ G94a}9dMΫ0FH1>v#Fp:9og(0Z~}'MTuccMz+ܯ_, Պjͱ0M$k)I0GhԑS%իQ87Iy w{jFetΧ=1"`?Q'W7Rs',:&%4DE,*KJkbl֌J1]X3҅Jttk/lI]l7W&0ԫ'֠rlЯO\c3!4I*$5*!^ !&)AZ0JfH@]2BMl2h "P.nNRiTA@bLP(krOIFШ\a}X1灼Y1F,jDVX#N#vq-^9& ? pDE2hU*%eQEl)k(!J A NmB,e=FE5ZLh ̇Y#jzQ%_g1.V/zwzӋk3)#JVrmb\a9O^jrXjrFD][oG+l~`q؀q}H吶` IF2["RUU_Kʐ!^.>\<yWភ?ےw_a {Voq :VaeAp}~MR~ 4{vfbe6eo*Fc6Xx>u!\0s,M'x~}~1MW뷼 Q>Sx`H3Y1+Jf\qXtܿٗ*(:r3:t8KmiQ" *j2O`2OuowWKR9JF-ECSn2YdզR?֬=lf;tTT^}UH i܄ ʖ[ bW>>^n^bN@,OZg!=^rPlIHmZr;H03 QO}&yYNdj3Dn59<2o:u?by ,3tn<ʠ5˘~V2_E3K`RGSJmΌ6> DcrBRgӌ2K6++yj.-Qh lL;5y"Lkp}0"2V/#hZ%"I'T+t[Y<(SB]?15śxqJ[3ǘ1TdO(5XM֫-yuѶe|Ǔ5kA5_6]҃ 4|_z|_Ύ|:{|_z|_z|Hz|_z|_z݃/ =BK|_z^. =BЃ/ =BD?*-ܞvݕ Pc\G08gxϸg@XP)LQ$-Do[ ;F5VТp 4 ѿx)@AΪ`+'4+,,B $&ޟ %v(&ƅoO~;'ץD+{}<0ٸrߐbe XU F &wg/_|e0Qg/{qsu#0 66VFs_~R߃ŏ;7Mihi̷ re[]gEA;pd JO矞¥挱tc>BPҝ{Н]@{)J>8\^lE=v '(ާZSW6?]qy1MEI8,C@(y$10xthaB D3sj$^$ G%㑅HD 2""&ZH0$Eh@KG%ƒHe ^0| 7R)"(!Ltc$!Lb(ܧ;┎S%T8 ĒO1E4֠Dp X$$.FC$ӂOgM] ^[);jSO }hC@_! 0H'l<Ș7NIÑkM+J (1<tv{*zzB^EAD q2A0= aGi"NvZv\TY^?OS |12 xp^Owe3.#᳟OSҖAIP?.BptNUQKfMRw¤AӾ Τ)*f$˝p%vqPK1Ff,cSmqŷEsZbRׂ~.`͡]7*+G+I $=2?.),ҺY= E 5)⴩w6eU`t|:yio_)Q\ZGI(A^sGq s<n<\>sӗL~/fsO_-;Tqxz~~Pr4i{zo{L3]{u44i2+NV8}fu4nr2z])1IUi>e<{~0yM_doګ`xuCY|kӟaQ W?I*TߪN57lu'\Zd#n1h1k = ̮477hcJKe։.c*#'ڢ,% ֘).ތǓb< W ay,0zS5~kJg^Wd׷e 4*7Zr5c S/ZMs]"E=fP'%iPPa ,5:yH3c\ ʢbEdǚLb#QHYzRaj3jXtLe4zl5HVHKDǚmgtaOC^}J罨-;(3v\Ӭox- /&Ak(@II L R;ECȮ|.,}1]%?ARCkIƠ$H~,1%/Q./$@Ϯ5;b̀ewR:G~_׺$oz)vk Dn$5M̚d?MG]*W;<(u4N jwe.bWJn^xC}a̓@ \u)֫WT\mt+r t)+lsl׃}cRU~nҚ^93)Du1Dgȍ`br%Q4b1v7Q{7QD 9ŴsD0KxJH ,Xeț% S Pu-Th8š hz̝vXв%4MXLAa<_iׅܸj9 ΦDG?.t+^W`JE뫟I_MfR`hR{P @9 X>Ne !y$! kY" o03u2ہF~]FΗ ݸSwjr[Ry諒P A2*%b+t@Y:޵6rBYuW_ˀs`H8@A_WjIew}DR&)`(b![QynbӻLf9 XͿMً7P̧,qt-˱8R&,ׇ`uл~:7~": &G*5Kٝ{鬌!>UF/YPt:P7:NR RYf(%jKΙl$ H & (a2bAXO_TF$RB2bҍhkSz$ɘB!}7I1DK1t,{L-auk3MdRv{dDBG@ȹJz\^KI^Je5ً<@g$rI*s&4Sv/l% oL ++6%cI0UgG;k%IRk']L 3Rg@-$^*%g!D[p"̘ Nqg@Eޛ@Mx|/?[wXϬֆs$%/$C #f'EYhˀ'x3oV5o8kA:!Mp1!zO!~rm$!Ie$4 R_8,UNSYb ȗ#2"(=D!I"+u>.()#iT칵*Som2"$85HMRk|əbRh{ʊgB hNr4\!>_+ })As=kO&hP3:b_q8W؊R2&lF2`cG{1ξ(*W mv{CխClJs?{6*qI!pKQHx?X9C!-lSfur%ytB5. lBu}>6'ǍPA GFGQMg;.UsWڛǦci,Yr',W}S;YՓGRB0d}ȤT41aN'0I@o㝡w'3MwT_e˱~18] p>9S|XnLwK%y;Thu֊F:<4,_‰zjj N, #;BI};KYIϚG )u9夑+_ P.Ee:ˬTX6RB 60&v` ] B"e!!mx3!ͼۂJNt fBQ÷DTY{S>3y-G0;?7~yv~&U?Kh-{4 ӡgvWC?g raeL\0Nc\Ptkfͧ]_8@~~zڍtٕkk~w$(}aI泼KV#yeABe4O?}7?|GO+_ۆFejǣ[ HYگn%5E]㫮 l׍͂wfoT Sc$yfM>;4'Ęt4'(0v4"p 3}/FR"f]Ƹ`4GB)N$C`0ǒҥlz\}amO7mqE zVZpDagQE,`2PrRD,0Lʮ<ۇ Ƚ$c=/PSafJ.- jf.t),,nnqKiw ezڮ܏^s\#cTQCN98BT.Rc ZdnwW!ņ)@X C-=̭ Eш$r0pesBEcPP;4*`)҂SVgupa<+.oGSv~|䦓S'~|e3/ "hs՝ .Ԋ̾S]Jx7!vUWH_/P@_֋O+RwRR f*8_7],d06.P5Q69A$zg.!Gvi/%,G H'@ @Zia.$!NDnXH`=(LY0GKI Yo9G.eYh29rCUymj!3Tnbʫ ɞqXԖ%9:J#a氾߭[QW>1Kq 믃>){ݐ)I ~&0wPZ"v ?&{K ~% #{z ~Jю);k>o?|7MA| Is1nRBu>C(/lpV6?B6:?Nk {_i4`£ds WQDU{Wbm1MXCjƥۜh2_;h.Kwxtl7|ܧe]ph3שdG%kG/+]YȍM}s%OmxCautwCw/w7ӻ7?^plom7iz_=/=y)F}F7ye̻<-ە~ $o}m:bKS<|}\مkYR7uY0 }#*lY\4DgtEfݐϥ3h.VFKȬ)&!f[TWDVq&amx\3!6zzN璏P z+%^-; ԂF"YLr Ѱl!fP&r%ytB5. lBu}>6'ǍPA GFs-8-x-zv\vjrf˵7k_&M1tנ;#E+Fy5Ŕ?ni^MsJ+!>j W*4;ݐ Z.0STD;*?ZJ$ǬwH.$fWJ-`?F] L9 !8<*A;ۉ & J9wb(5Fk`eXzWN:)= cL5gV$9yp@./BVs/} ][o$+F^rPDQg9 A $R϶^=3Pmo{j)sqWuUE~HQd՘C"Gb3y'O_QNZVmThn|drYzCVz_m]ƺGVCQ;]2R6GŠQkFct]E{aq0zTͣn{_6)^]hI#6+ ,2F%2B5FONcC"ban,x=[B?v6V'+#LY X1FNtlj],.*-:`+<"<*J$sJtnhgsMf׎MRT9 T ϚDN8xl%%4U/虬N^ $* eУ 4oE/z~aZg2ѵ҃b H&4qA|ʩU)>|V7U&4IWƻeɠ/83evB ȌQU$mP(Su270yKM#ptZ.?΀Wx s,1u"jqٌFT,"Fs f-\jGkOyh]E϶ȺƇo eE'4r6uPVMou49f#Z\c,1Y$$&/lʢ\ 6d,Xkz1>YFCۿ; AC?;R]g}W}BϾuaV \iMMN#P?٢}Wj-/, _XNFgl+?ڪ,-|y lG=ZVȎ^io,rc:lܐz[9qH6Thcu(n%)Ν -SH\΍m:Z聺jV6y7^w;2<$ُ}c!g.i_9d/T4DWhR/Ң)r,)P=I^:om>ao¼ xo{?>~-qXX!(dl N+%ٱ}nAe y*\7Q$JeW N(jNlM}c5HSD{VHggm u*9M`鱀[N?,)p20{<]1`za:,W|.o_|܌|~W9E]kLUUFT m =qLBt455˻K2oH2ϵFsrpᷣBiO`9<9k3D 8ڃ^ yD7@fSAU[xL=qjjw ū=3x5"$1:BL U,su4T\>QH+-G67j0ƦbbbRVcE*Vh !R[z271&. m C76_!-o|ft4B*O#y!cAƨ'1)*l6S֤Bb]ΰ+VjT)(y'*AE n#NPtab$w˖_az^Pct/ZʧxB:U%<:%<'xR~駓ePҠ{8/:y6\TxJ-xP%_- ]Q*bkJo$eN)5e09bhBuPh퇳]_j*! %`YJLVLi'K[LU@72vqdUaa7X V/aapXx%Qƌ]=^iiV./Ndrqr8;{ `R^@H,O i+xt^[ImHUgSQCs298!yv* *>%㈝t[簌0XA8Iɱ-6rBP%O5/^ĉC+&!T]nZSWcNFX+juX8hoÙRΣlXqL\cPxҀ$q; ЎML,8.kbݴd]\θhF\q!„jGp`|h[ȿftPʰ6PAxQ$űu3f#.niǺxh;7aJK:>h>Gh Z! n~|ǀُ hfħ{m=zlՖ`і^,;cVkqP_v.v+<侃:;2 Fmn(O>vycSN*NRqNާE_\>9,/1ozӳ&cI;S8QʟvNJ.pQ':;^)'~ûCo?^Cx߽e<3.2nlvh[h-ryT+U!$X,[+B};OtK|T 2ն))T9L $,-uP)ysp`3aI~s7̭6ƴֻ<mu B9$jiݵ1 xP 51:ud[:ILvL$U-c՝^T^s[.=jLyP[잺sa:ܠYwr)CꎳE1!aC,p-4}֔o{/:pu*~gп>[I<S>_mʘ@\6B[rfhҗ!o 6UdSq̅}L.#,{8Tra[mٻٵy/{O}-[fgd Y1Z@AP+jђčRS֕<#"l.mb>  >P!):ΥZL0 Yg,bM]m39͔@ |gI=-9"3a`gWce.U.HԈ:#3Ǩ)1!gZ{8_!i0z?$8F2ø5M(e C%KqH$8* CsѻhHiX!zOwO`v~֋b$f? u#P45;L ā<70r"4E^)|??yr%!D/N :ׅt@sɴ=<ߦ!ktq?k*eٺgv^v;mb1+Ǟk2*YgT08CG'r~T_zSpBo pҞ7C3٫~ʖbMoN>l6d}ѯ{7ߟj&֘=E1 {9fzacDo`p[4p{lZk}HOBS quk׋wDGJhWbxb|ެ,^}V_@(qŋi}o?E98WZ o^i?m{7^g/}?OڞiOvޜUs(~nZ>;Y.h||׿|W˃t)GՏ߶|hq?{y쓃́?}|<cj>]wξ~}|/ܼY-|v 0>{qsȏ;#WKj(׶pvTn?ڰ=׎F|>j>D:DAɼ)snQc|V. 'dM~CO/Ǻ,: =3uBW-_[CS-:}Y0`j}Ȳye~\\10KWل5m¢59zq؛s^<]OYglzEd`dGw=nOcw}F˕LՏ_Xpδ6;l3%}&3ATN/.<㸽3.e#%E E)L;j&UG^'PlJZJљdb~y:ip|nkܜ%T!WT 1؈,-'"#7J7Ic2O݌;ܑ͸Ņw>W!J~>+?Z(W^7ǰdZMŠRJ 97< JlzTo:ޕ niHMKJ&+1$W4k-ڔJڼ%ln8JkKYGE4ko>fk9clV"$c!7;/tNf~t =RBkWL/缶 {ՇYkǵvʇYi}[Gje-Cblǿ =x_xaX/p"JC{GЋo v5ayi{[6Cræn{]O?fy6l͍ڼ;[폈ϜY`קOfag ff85Rx47.N{O"kp?~[ {CЊc}"+F&띀&ݤ3-|wJLy՗AFz wBGd{sN._0։oϩ~cʨ#߮A=~ew~ ,^j&jf \qs.]'K-u;4FHEqDrnszbXFn:يpl.?xhr͜dS$9rfLRLh!lZBVK8-RI;m.tY6x+߼f_]r(g4îkW}~mcg=?|Ssy F͕hjhnEd*>hgKVo$mTbc;kCI>DP(1y!TjzJ\X HӼ(m\hlUi|?ݳK51HqSZkZ*%J͈Ҙ*.DߤnERK5EOʍZ :Fmny5֬%)|BJMQV,N@Rzm-F$\Kr58`mlJ9 ol6)J3dE D$/b77aGK8F|W:\,cRBTǵ; ۼm&+Qޤ C(R2[ #׹=&xB7TZ <pfۯD#!^D?=.Fl2izETjϛ9鑬c'dɘȺUcRSu.^?o 9f*6oD$,ۢDsIEH `¾QG[%c2(mRt5J !IU 5L(( E0J!5ᆴ,hjivJcZRcOI=i]:g("D OQd}Y D.FL>yn9* @ #+l %D: x;U |.%ɆaJGLxg@nj ָh)4"٭(Xd5r `-iЮynΊ:8U& 'u$jȤZ,P|պ %, Z4.N* 2XR S{Х=Ť<$J/Z$җ)4C%lD) e@W PJNZ%l3 ر-Px hm%0: lr=K(w,npZ[2/@8 pdI..x]qcQm\g$0SHO&#2~\ qwq`DʟxX+PaRBM}LpY%,J@'2В-КOdi^aIGYӐ`F+flԔb譈"XmX@hP&%lP>oߦ\%La R-4n}]{ykbp^E͇Ǵś͹ Y* @z qt*MN"i`t:603î)87,,:V +eZ)DZ&s(Y<0Fr0,u:~7lDŒH"ျ!(A/Qkt CB1Ts \NZ-Hד%\j:R ;( , 4*DŽ*m )7!0ւM[ [.[dv5`+BqҔ֕& xrUpՍ9%I3 ap`( ep#IX(ÎW=U4+~(]ڈr1QԠ80*`cGj҂cMZ b'^Q1W0-lZ!9ty9>NKiB+Mx7[E0%V$NHL)#)%ǸF kuJ X] )lt6t%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PW}J R9~@F%uw_ +CW}J +++++++++++++++++++++++>_%(I ?J QJs@rg@Zw%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PWu%PW}J 8)mWū")@qmѤ (jt%%oH}XIXNF13!a GS dM%G̒C%߰cp*Jh*Jh*Jh*Jh*Jh*Jh*Jh*Jh*Jh*Jh*Jh*>@SbܘZS}jRMͪmzn'W _*/-ti":"ڲeʢH&f~ N.RefԉLq"U#]Y:Ig mkR$2?s-?G] yyk1@fb`gf{7 U* CN) uiEUUZD@\Ah=;0> 5r 8vM>-L@>Ld,vك1q-Ic]xC3,ȍU! ],Ф>x9V2@g]Ch~䖨rQK0kݏ q8Po{5bϙ4K.X!Pu eac.&`4|;UBOԝ^|~ t:킠 Qˆ0zBJF!V.81i1nw$HnmVfr1(pXL6QuQ0yJo+,c X~6H H~,m ڬ&|f;y>[GTs@fԏJ<>_Б#1{4^2[-ƣѸ2AG `'sv+` 1:Ct>#Ev6fv]Aw^ZX-->33oK2 f=(f|#ےGgmIـhHGCFSės 9"0K-~Y-˥@//\|>\&nEB(*rc J "p>M\?T;O,*.IHg8oW_X 9"] 6!)#‹^wny:VzJtDMH.72T*ièInۿڕ^s6¹@ #!m֡N&M8fM;^Xܵ9n }ҺԐ nEؖh>z-2Dwp"ubo>zs6. 1{8{ N>މg߽a=<|G| 54܎/|mCf0`(Y*(?nJ3 d߯MJiuLyuByeo:,]ESf\A2ib<"r8fakyyyX+!'"/γHBd~u2in?jޛ:1Zs2_,(f^f*ب攠\w߫z/zccF\{2%v1+3r?:+K-& -v,мkO!牛kCOb4qxiH\"cM dq($02n&LA4NO0BkLR0(zQJo.0sWXHi;OK /2Kc Q^P ت14+eyQ|-\jef ފafHT @tf Qbb(MݛWIp Z(t.ONH@, U $%G(=)(վ\\NHk6N܋mˏp?괲[wb.Omz,W{CDub8⡈p@@Pͼ.9x@J5ua$p(a15M "Y8B E$@#KŔb4r`~&7aPB2dṋBQ9 yD6Jv!N'`-*R\eʩ8I6oT6E!gk/( dhLTd 撴EsG eml\zA8?%vL#{LHdlsp;PY齺9=0Z5=SVTͲĆ:tpr@"}4bODi4HOBzW 8҂m9a.6dDPy$v 鴌ckjwHOʙ +J*(dտUYb:;l?r) Nfmlֽ^7 |Y\]q*kS Ok+Z;essG jT7ymݞ /iv*.4$Kz Zzay\ h':}FZ3Ruʑ([=t0d0I+i1ifeTVMrdXLrT^Q׭.))r153:fsuȵ)n7Yv*oK1))tʁ&ޗ%iڑߏ^?_v&/Ggǿ=@(9MN݃&u=ƫk:w>fr֡MdY-yީzqu@P)@\\k:dܙ~1cnKa''w~;LNލ ͝Iq,R+n0u~,ZWн-  V12' <0"cXSXJ"(XbPi "@4SٯL8^OYLu2{ X=JxIȉD8P`@=[ȽM74^^Y/cgc1IՆe zuC%Om9lg/T+)c vO\wIn)qζBsa\s?'X)T1`z11QCfT +m $cU>8֞ ~A"D"Zx `@JBE,chDŽJd҈GT[>S-cѫ`>=.A,)SM,aL"GFp)XP,$Y06͂!TOoWoMd?սY-p[Ek֓^fq(ԧKBULB!+*!ň X"(ۀ ͋s5ggFR)/ "UIJ `lu6#5.6}r3,ŏ7x2{m8? 7CU)$a0`gBmê#zWjf_/ m2jFa=0.Nk-뫔 3sƌڌ팚~6L֐qzq/fdWGGrv;^_S5|qum=y9K-ⴞrk 8[{_J5Ry{,LʑJB*()8 !kZ5Z$́Tr\4Z#4!?I`S;mkR$ɕds20j-]2*da,\h' A~`UUEk‡M5>{fgDπ^%9«odx{d6*hCO/_UT~=d { *}o OhYq3[V(GR2U\KP"yc"]EQ* DOɐuX)J9e ¡^41lFNFE**LΘ9ˤ * 2Z&NҽL&TR/C8d;uUq+>N/r=>vnNt~ɪ&?7 ]KL/o]7\LgݛN _~{/:pE'I~5=ތ{hW/u~zU)y{v^~c8ϋ/Q>'h/|&SYJ,ŝ?I7޸閷ieT&ƮJj &fǿ[ƒZ]ݯo-D~,.Y?ZRŏ_\"2t(< `yK,ۅ67/դ25+Fv*)/&#Y(R=m*+|^ہWϗˋ, = Z1I^u֮2,OjXz-[cMM_1ָMsEoU@WBHIX'!yiD&&MFʶ&m[ QiZILH -?5T&eǖ1{H˒Hcξ`?`oSDdHR6(Y@jzK)*{2@b̂>P~_ue[uZ۬ Q9@jG %BNxȉNRz+;ո1^~D=n.,z{MVppnN/Oȯi$94AˉL|.7G[p^c2)礬%g!IEW1xpI8ҦZVzV)'eiVj;2u7'y] /j${i^Qgäu~l{dzCX[ip}n揵b% 4C u,1:+{Qz@/oR']N`If+%ד7)}ɐʵt>CZ~F?ew}Z"hM@OG@.ȗ( B$1 x0+r`F@V/Kt]- qL-hfqj\dbOE^;0}W zɘj[kr5bjF0;fqy@U$1حN$_]Qʿd4k@w{Ro/EցN; UPsbhOi|o(]j>8. 1Y+QwZ!ďd$tIHĿ.B3]IUQ5~ucyfX˸ga8僯+K3atyC늺M<ك"#4M_%Nd=sv HPFNAUzF1/Hk+f=Gos2lt5f{ 9Naحϟv~0/| Ϙa6noyuwU{.&2[l@z#t5\|Ok"JF;B6pҶ/[;/Mro'k b쇛@+aj#`b.'m{ YAs=h{k`f9%AR% D0y-As]Vbvz o_Yyjbd>Y=Lg/|D2X *;VOˢ$dyHVkp2e^cxO9+$j_,Cq\$歩DѺY3rs8+z.WW׾X:薛Os7zYfYǷPnf.4skOZ_Me{ Qiami.y:unqܚ=skun$I/E%y1B 3kdXgry^o :sﴙM5a]i,$%:8ؒ(Y HMj\zhB}ԭg}޻⺘jFFxd3v5HlDζo7Mb8SJΦ )Q{z<%jϞYL; WgGJEN>ؒ SePբ6:E)q"v \(emvW,lsko.йxrF8sM jefvY{\g}ٶMGܚ Gk2%C8߄\Z!i5н)Zh0ŭbGMDĦQ? DC[H+&]m6RgZC&'Ә8綌R17Qzu*kJU%7ڤ^G$w,h@*djŧ|V]n)Y-qo7ye,l ;[$P[:Ĺ I *2!.ufSO4u-\˿hů[Y"q4g+g4,IXΩEK,rijh궖;Nx~b 게RP(N]I @(kN̔"YN{Y^|rE66EX1ɱќw.=7v687]}$3K?x}4[Y#.U%;cl[O%ؒ` lZ.Qݚ&QO?2>ŘrUzqJj 8"e^IPk6U?oyҖ'=S)$i8Doe Bl bu8.M't (#m$\ՌAH:4w9T]BtM6_F}MEnk|4񭖺ŪiəȣN$qȾq[3Yq1Pz\ئg40vRkd#Vr̥!0uctXח#⑓"`=H' &؄9vtJեΩW VB1|g/¦{oEuxSCnӾf`W?L'W {z8xŦE.F:;Dja="ɐ/]56ZrC/$O\7\Ǽr[ He ʋKN?S>ʔMixWhK2V.m;%ԩ1yR&bF$Vor K@=A[Z>pXqU/9Y_ȏܛ PM^'iPnjMFH-(.r&)ܷ;.3)l͂Y}vv=Gp.[~H&Gѽ񛓃6s4}N.<i)'Lǿ=e5f5v(C/zVsEvhqdKh.9s1d:mK%rqk6(7d>z.LAROjdcBuZs}_b2 fz`t m B2!M45XiD/Ga@йs.[H'yޠO+ߡRf3x_9)gO{ĥi[uDit'yv=jՐ̆bq~@њojϱq"YːYFZgqksY/i=b\$ S2$)h+D+ȄI@'@9hB˷hk;]RiL1@F 1b, h* 6j ;~o0Ź,󘑻 "=tffl Z kAO헷O-:# StTqYo>I9_Q@;6)Fs-X|tTR#\7v-9cB?}!uDj OC{c!Z4*O3{}{>/utL=4w?8xRg/2G;K:^R׿kruyYjN 7S//-q ݬ8`*:MKB_p=5/|5&eA\i1$-ѕi2 ſvk}G^+.Ga{&p .o~s&nݤe_iWy;IN'[vD+/9qN(ʫ]ݗǰ~%E‡J28^>z 89nf^ýӴP",-i^p|iW\q'z@og>|<|wt_;񂦏 hN'vڸQ2R@Ý%&uޟ-FAC'gHZ:YoѸò?')z#L44?P/{YZޗ:]vw7g-/1 /^8gHXϱ1󟻘._j8OPH]^:K->֪\Q]Jyuǜ޷5vp;8ӚG)Y4vG,'ڄ8{rWo%FaވZzj'y\mozp<#,z[pH?8ii d]_WR[acc%AnJFZ>(uO63d[`mwfn$"dzXER76\HE8t3kԓi۾%umhd~JKq&բ udɖUFK Dj8u;J;.} qTX+-6q*r=x㻱z~.Bm|L=kMQlW! ^&⠖әaP-u/MoK׉b1csB1{7^ YTfUJ11ǵ;pF[(31|5M-A=3J mp'*xrøH\c`7i5"T@ê/z>"C5~g9ˉ#8Akz6Lto~T=dS.պZ?H)`-iR^#iV5@ThC] Qb]e8ţ 6lϱ:aa :7 ]K -n\oOiR3iTeB`Tk5!!(ͫѰ.byͰ*4CShڀ$mh/J/u{99'-ȚCy꠩ొK*$# +h k[S9U\*2yWn@B]샅DՀVi4!9cԥfGn p @TBJJVL t(S% |U7uD&ݎ)rY@rMEwJ0!fZT!Ѭ=K&y3 0#|SS8() ir>,UL^뺂x2Q2Ra8P)U:/֚SkԠB'(K Y'Fp|1J`Nւð?*Н vҙ0GcjmX_W\t[S.xbQ6=3, 0(~i0Бmd%KLjjc(Ov1!6$/UXPh^HڤJ HHDLk* !yH 3"v̪FGA^r'TO\dͶxn}[q;"CpLKW VfES ԏiNٻ涍,WXN *}UqpiA/vZ@m .H_BT!BVB]J(v1tm6Q*B#`q20j`< &cU"mrFcЈXQ@mU(8p:)IX@9vi7F:;M 31W&J |ʀjw72ZDpelN`,*vJÜ+xrxɁBaA1VG[EviXfHwڲh2qYTJ ^sD}FiTbT[d%^QEXYm4+;-wKPȤUKeiIF aK/t,NCֹ)6kiLp ,+5+5l:)9פ$I%2h MWZpd3 3JoS9 o)wB1"Zh<5w*(s!-8Z;@Ƹށ)G9Z~9#} _Gj7tJKo*PcP/Qnx2pQD[NQ.Fw#XP(yT!J_`Hz;+1Va Y!|oLW b%ϽWK WJGj ;u6ae èʋBTFJUJJňCoQ?;B@2!c6!*ѳ58)znf@jX; |Y 0x >@$40ҵ,r7n\+HSyzH*v$kJ)-XBaN;XriYɈB5gxQ(+s iNč@Fk|T;=wq$zcvNb6y$rUQzd":B0:"JtRcEH oN](4\ 9z L(\cC:LU~ޮH 坊Xdgt0P5aKB&v-n1WD@\|YD :c0,ɕ5Q^x,*U2B9EcE)mϥ|^߫-ݳ"47^~J IF $J X )D%I DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@r@YHp@0ײ(`wWڒ@_1H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@@͆Bl8J +`@i^ +"%L>@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJGR H vQ%s܎+RKTDI DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@r@w;'[hb&ł_Pq֬dD-=bz~v쓬f$w5. 3Q)דx0lbVIv:No9 ?A ? ,t]˿yZ={pcz<_Mqt'u6^_2iz>Q/+~y."wpٽވKƗJIHi |3>Opw!(Ɛ1+/37Yy@*A¦c3a;7 7eMP2 ͫpW4qcNYB6X_ ܬTM*_eiEs/@b">G.RCYUV|JU-VQS7qe}@TK.[@%lZ>&%iuQ'g#<`f(_!9e3S Y%Ev$KFrN*!rV2@ZoN Ӂ|j<}|%sMӬߵDv큗99:k>"_\B(v{M ;0a %LC?Y{=tsfP*Z&&Q3>7 Z9B)B⡢W=/.R&oj/O d~ķ/K.TԻmGqfwpÏ6쪏Bo#Fak'cj6Z҉gg?#ƎV'lvw&z2;5K#=8 :=Mv{Kg-dg{PjkvU$]'ZMQ+״E;ȑ 0{wFβ%q?|(I28N˺Tvi\Zw+'s-%VκSMwe;Lm)>1>Y fQZβ+Fޗoy< ޘI\WBiѐY:`RR2TͨI!߃55JpY'~,Z25 yXlPʱOLL?hHB̬Ѣǩ޾Wh_?ǃ+Ńth5(N}>.ٳ)ZB:݃t:GԤKˌaҲ-e۠Y@c_Tn^ nd+ Vi_1.[댥e;޽'$~+,IJp9k*MvU!m=t6 (jS:IK:L©~#*jPj63#h5zx)W{N<47E_ #E):횩ךjTk*6\sG#;-fkwwmFiyb:W侙5]-Z̍nr=Eug<-Q%ZN5{ g _d" K !Od9+?hx۔dAbeSS ;WsK'&{5loƋw45{kZv_4=tz9ӧQ0x_ uv+7@;3uFELΊ7߶p3{cK0_Weu^4ơzz7j+J'u/!lowuR lҮl_9tsj6Rf@: ӋINA*]gVx\<>=(of棗~9}bru##|wCf/m(bIUJ\D-JG_d7>/ǭ3GmwOmЋ%}Y!`T1t&(] b'AqEdO_eY|,sSɃ>QRm]ȂA!_|u^{L ʶ$b0Rk=Y.u/>W.@M `0RJǭ z/>j_gsvFC,Z:Y M[T}MD9ۑgsH&A/-(򋖇\ ֈuwhyu&QkNTLQ` A7z|hyus e0~'zRZu,ԁV#B45g⁑Blr`T7nju-$j2a(.(CX[J[E߂Tr[Eiϩq'@jRuO0oͩ^DK^wz0~7h;AV3A1VэZ|׽XE{eĬ5\k`ng6m#cׯ\F]n agWլD]7DS*X8!]6~jR~Jƽmo1oj%+|ꠍ-"?yGZ` -ڲ -z{εg2Sn-Mײ%Ze vOvAG.)u~&C:?wu,';ډ fmٺ2rk`K'T j/0 k>?Br)G̟\QDMEY2e؍+7J,N$Va XPm'5/JE=ȝWk}y=ppֱą^`b?\o?@[Phz?LS'u #1M&;ϿΉ) n 1{Cp<\%NOKz/q TaB9 `#_Jkjq'HThU0ѳjlmx$X?btet<0ǰ:-;8+(V䇥~r-B:Rf I59 f8AE}R涢%WLˆn2O2qF`ʱ]hH=(*dϙ8XBA@khg?sߓ4Ki2z,a+iSVJZoj 쭾3֜S7. H,R*=F7MɧkdLsfe=QBX Z^oqf_NSZFN35Y>a¹Mr6;,Eͤq"3|0Fq<Kil}ٻM R% mf~" @) kalX 7[o1=y} Ƒߣ@rcڛkm%F8Z,9J:LRzP`[}‹Fq^/WA^-7IUp!Eٴ9C脻8X'rrOȁRvi*0ڽn"ZD1MS`dXrA*=I)e8lۏ/]{Gn/Gꀑ$"/D!D9ĄG\]{%.9,V *76">eBp݅:[N@@ !$ë$U17Q |Ej=4ԪcJ:GrbAe> @)L %ˎTUTt#3F @!pcvb,/K. ǘ'Lh$)S@J0%vb@# X,)I]V1KBj%I$TYrִ:[I(v0HT7s{7w"1D;7둄)*zC$iTaEW~h0m)yd!,vlrMWXM;{*Z^i M}RSdTyG, Y w"Z`WDytboaFo;ZzIkz}vPYo9vJ5יLQ0:9TNeX5S9?0shJ#L+x9iCԖA=Fq|o8ezv&4ptX‚] neynK+_'٬duqnEiE[W_c?0}skot`0_TQN~%00( KHZ/ jU]_?|)-Fd.0ʫ␕,MghnqSp@<B&HY{ύ\`ﳪq{9,$03iPb8)Ɣ6ϮKtSjEX0Ɗ]c8f9=]I2k] EA S(ERrXXqugیlS`y6*@C\^hAn@[!FV!"tpWQ~f͸7Rnlt#xjϨE#'=B ޡ.SFjc`v$w%Z ܞ_tӇؚiEXέِ$%,KQ"@iYiЇz6̓P]%㘓[$UrTЈq+׳ryeFczum^3'W{9a%g3/$) `PʰhBDB1v.M{_7Jꈺj6]䳅K7_dL|Qp&խ {/>7kFViAqX=*8x:\ Z 18ɸ~'Q'D#ĩ$(PVHЅZ#uÿ͐-hm}B ]ǷG|tڨ 9;wVP㒭 Mp"m㵗[~HΖF1X#Q`f UΌ:5h\cgkܧ\1Q!6&nR N{qh}^d6[CJn; Fkoumԝh^=ݗ_Xȧ8vh٭׹z3yG3!f/'e, j]uY}ywcTVf$1y6:.>-\1.ܝ'LrK36 c|ݓLl٪yˮn5l#]7,.$1#΢ZB:@OVcx鬏ubh7&3);opc2f.ez=N$SJ z]yά u*3j\k)/xF&ww=ViR%N $a U2"]i6SiJT!.2LHenQbQ&OIoe!Y'Uz*ُڲ]} )?T|2Md'. T897Up{*]Esd1s)@Yeԇ8598>_.~>;>-NZ c|DXt? zץI_. hDEH|ƌjuPP'Β 3鵑1.[ eK ]N̜ fk} T;3.8&Mo:x6.w5BVF Ju#Z)G-NAR)]P O LAk4 ɜ-6Dp!4`~˙kPZҮD_čl/z; qQ5}x'ra P't&Yp5 2hʝ0ɨ3ET !HΞ1O9BL _.}fRHttrda$3"@$%Q*2hn=u80ili.]?^ON/kovGリ$eJd]THY<9ށÿX[. nĶ $;v2~@n*8ʮ_%F!땳YU #u&GN$LE )vu<,E4Po(`XSu-cQhX܍ gc-摼R\aeMiCq4=C1o\<*ܲ,E|jB1e(`]F"exjFNWݥ9~(), H lтbOӟ]hJ=J"02brЊ< ~X5i?E +)=FI6VX!D9%'qfB+&՚"#2?8E}=D; s"KʳV:s2A#R-.(J )k$@M٩vڿs"Pg8+&؁šӳ܁nةǔ8GȊ![GJ?YA" x pyCӡ>A_I}.9(G¤=*1#+N\(>۹&+8(i;$ $>p: 䵲\]nqp>J{CƃǛ=2c:xv"|aΈApAq9?{(Y& (lɒ6@J0JGfKq8g{5x^^Rl) o/A3#ЦMoEAoHg m2MFa,'eX5^7N*dr+1uF_V}}~r Wn.x}+ʹ"(%JN3۲+R!p{utHR0a9LҮibw6s]nB~x4^LLO,q*d`ØOHoy"Ff_(Q@8 R;ߤxZ{.z!"J@t5)RqY1ȲPh3}+?YQe9:mzT`]1\# $bF"`E`)DZVڛS4 Pm)e09 OհiUx8ilf0ewWԴY JwkI"&@JϿT65?4eޝbhմy6Y4yN1giuoG-'*z5]nI&JǕ&\Zx󆧩> v2V3.#r3R㦝zrZ5n4 6M_?%]/TwxHVx g*b); $dG:SI#89lbDlAV/ xnP=CM@L-e Kh䝪5l&S'Vi ypT3y#C/ň!ꬅzM#4]rd-[1' dNҮB1e {L_rԴ0Z?Ic ^TLؒ߰A& gҳv4m{XRKWnku[DοLCZ:pG]PΓ,~fhղ7qz@ ]UqN7&UQ!c˴"CW^{"QAꔼޠ5{.T`/ܣ'\8x&J.7*A^<d^Կ[T/x[h8i7/+]QpIˣTx}u #:*>GXbTi4"cq0ta`2,_bbPzhLny疉 Wwp*u1w<9n>Nd4"v/>I>'A֋b${9~&YFӺ\e#iMsO3ӄ_cuNbUL3sKjb8 Nwqck!)`zXT)%H \LٯKQ.Vi6* 4!VmLCϨZ&Қ6+97u(}H\Rxem=Z:JED Q)rp7,Oum_+^C-~_E;;Z-/㼛dڷ[OM9]d/>dG8 yŎ;;y퇱x WpTWsX. \f"HX⸕D'0}XTA7CӺF±Xjd~吪3PۦN#O֩MLeqJlE`. ,vO+ x7weM!Dv NDMԲ,™^{eމmJd`R4 +xVx{.F^/mʹ$!1/B}FK/I~R7aCB[-)2;bAc1=:[ >|Pz_I8,ǧz5!w 1̥D(!DšZ8s䬘cxbmL޴1E 3ײD=|f+q"Dٽ)/h9_<߾M.y*mxL"f_TOY!m ajo}wt(KLt6E((zмk1wC쑿LN  {?]0rN;N_"÷ذ春fB 88k>u~z]#IWtqa~BRPFaO:qըjc!*dLLG( :U52mV/͆">TȘg-nf QJȐW:IVمC**d Z!l"ӇWw$׍"j#Ph{8po˲=]kQwPÍ`) rk1(q|UiXtZ\̟^CLJ\t)kLUftek  iO;_};~Tg>7aHr  ʕl?ak!;Mw#[_ǹҐS?Gf_y!T5)2l2lS*rH?\5ń,AmmMƵ.b, 2NlUImnb C. :00yc36Pl(׷FSW巯53^(?B1{™L_zƻ!7A8`GC&n cT) B$+'q4ZzSځ`lzU8xnowx_ zyZA`haM&PC e: fIƊeٖ`UeZȦ=MUzuU(dm#FԎ0&Uӷ< XY§mIt;559j쩓N%e}RmIwHtnM7]7kE+pAgyW B!{otۏA')%3d_I0!iS}ꉁ7O@Y _83ތ=uzE9Ww8TmiŌ3Q6DC jA^0yFΔ`rNz4(săPPC1BOi ws⧦_'iʅtNsp1SCfapWL핪'"`6HY .hS&?,i$I86ۿ=肻F&mTco]8_(@VK^T^A3/ VQ"RX/8K>h7дUsp(Knz>93;cӁx@C*GLM]j$2hDh_8%T:5qbGTTP=t8jR7*5H \xesJcac, Nse d%'WRpƩӶ$ Z4|X(+r';Dж(t`XK6<~͒xt߷6Nz(He`/_08]]y|#餳i?)b;J? ]k;7S}ĭ&S#>0 exL5h(^d}X Ǹ?W+*5Kej]ؼ eF1vup^Ӏ|hkMm4wv h iN`GCUeٱp^vUFrFi# $(n 129y6|zN]ẀFjt3_Q6dc$|/?t~QU70%ZJ}bgS$b~۠csXb)8ﮊ* w߆In2W뻙/w =~ ZhZ :EUuD<Oyс݇nҼP`Y\ w.Z ƪ;PHOU6,a&0wQ%Vת|w4O_长4ń-kl:* ޥԲ-$H1O;n;Ե"ڿJPEPT>c`>n#o)@Ә{WPKyƷ@~֩!~^֗e^DnO_> K /{| 8asG _U{Re\lRuQpuYmmӟ嵬vE~y.krb(J laczZ x`6|!w(ui,̎`ެoר,Ђ ]ТOFrM +$1Bџ(?Yj -.|瑢jǷfhl[QfS O>. kk-j奮v3ԵO o2= evP %} j~"zZ:c 4ǽ^00 TǪO 'c7pЃdPjMƧ_nx4[}F'=7;uW괿:HLGK-ɲt!TDj=-۠A_2)m"yuiFy!*]y ϓ\>imDw=nVN9%n&_`gXt xb^*jFe^U_ K)s <*kήbO2YS'_n><\<&K&`gXt 湦Q0/ʡZ8*<5:7} Y*991ωh9AQwS4=w/Ó9 '#waԦ%F7fO)t$/Duz4ذf0+eJ.:?F@Nx<Dӏ'J{O "$[9ͽrc} "(\t =ۇyu >wrrOA 0~kF[P R )%rzDM[JMY[FgB&" 3ΥIʈ\IeNr6܁P>\Mu0[?-RHXSo*XT*b>6$3Q("9:`3J4ʀ`4דڎD&SN'axM gv `Ԥ6>6@y(MhD'vs$bc̋.94vV']#GЅV$X[IjPwV%(U<)^ qy<34 ]w#+zijOZ;ppS9w-k)j8sMinAE4ke+eA.y(f|?fw.+Na_JKS;SjxzC HdšViZ/x+pjc {Y)`H>lu-\2 kӝ 8>ŨTLѯ?~|Sт2 #0dfGfxtI~RC}=zIBaMf2Sρ- ii8*k\h ).ӲP!a5?.$|nlTB*C :Xe|$ wQugqu1睁- ˊ\0hBiJg~ D!v/t (ͤ&F1&sL3Fј2)*42& 0Ԃ92&'Ӽ;"zނօ2EҢmbDgnpxI}¤[/М92DTFq-vk)W0 vo zqN7sSy"E)xAO i`?=(ҡ$CEN^ M9i7+VRMuzvɦ1HD7P45 k'~u?_qswߜG2|&8?Y3g9*M28מ+IZ6y^fީk.S{k ZaNg&i,x11JUVJiKc \ilxQ]kyc36~:M%Hr/u٠f3㙜SKi][P(#p@, ;ʭ*l{iIlhII) xtm݉\=;mWv˜SԽ=;?_ԬjQl-<*֡8,fW`lUס@%R:rU~\H\]yvR 0R4+.}=ZւuLi|zhqɖ-(ͭ(4'3r 3%vo`l8RIԨR‚mG:@ԹuuY2_N4•gpR\[M3BzDغc xya@^@݆1b+e C.їZ8}Wl/6$:ėJr yp:PA9/EA-ʜtĠxEJCx kk`\?=ᗍ)'T/ލ *g!H]hzgIwxrQ-8e5k(7\_K+m#INc6~uX;/k)EJIY7!Ȫd:nX f|qeD$1Z߫j}79Fw8x1 MthJ@۵4Mv(pJ=N) O[6Zx&UhWLqtO&knz[4Y 6Z/o+;?{S4%-ޭV¹ &Tv-Se}1׽Ax` X.tvUgϒn:/|Q!p{~ިWr::I`|WN ߵ>wr2~1\t //vz3# I4]̰"V!q2,16؛j pr S|k:LQmf6Lp4B|&HJfEq "}gһF }TR^t?Ɗo7t]UǸ׃it;s"G !oH]&3H( MG9k鱢[H bi8vZ{o E)t8JtfcA:Ŕ8CZ )I-]ZOe.hT\ :$6DEr0 8jz \W-ob֯zFRX<=CH `쀹Y(X6b9z^G ==oUh==XvSzuwq%6?8= 6ވ\rb3L>WcÃCqg5r8j:Ob^nHĎӐ z@.OחDƵ=tئ=G}V}љ7!+_xnOE(j"(/L"F)i֋;4uHҗN`d}`f=ᷕj# OovPG$<4Zaй?,,FC8c*1{YP-؞,k<۪Л'u ]N]-_<[IF(3 W(1J?yq34v@n5Kh]{qZo" 6Zc2YdzG%R:aۿ&dٟVז=*| ßޯ hyÆh{ uGƟ_z^~1ng+˓vJ^];zۖ17WXS:櫋c 4R0|!`uy ħ#ii^6`7_xNRjg#1cvذLyi[mc̷bc ;D+)CPgCk١%=%wWߘG[Tآ=p/ӊ-әoI*骮dQ28+pJZ9S;y3{k~_fnH0#:Bd[ |kwPTz+PoƊyuҡޒ$_hk9noK֖^/He_m :s.ފL֒#'YWLxYҠskm:cYtM ʠP̺lb́lK?JFCe,K zfˡK7(\vwiBYEuG֢uah߾=bc2qS2ź4Mepf-HAUeY=dgL,o!PBc,j>(eNZ3ҕ:ljh"Jyi8I;M>^ī<4PpU!z[)1w3"j%oHuFA޳p241#NF@^5mQCӶCAj .xzla؟d:TM x9a`E\I*ZR(۸/dM%[ 7 |ieO!%I:/p:Dʵy۞Ti 1-CQU 5 H)| NچL䶃iZRcpWYҒ 1̊ Bo [j a\ҕ:$l2}mGᥨ]4m蔶ּf /}"+F{ߙip$r|3nLL|o c>  /"|=֏!'Ƅ/ skquh=`ИeU)erЕK%?b_*)5咟r&1!eV702i~'WL0tB1QJ˪kܱ]F F7z(wo,A#5i2Ϥ IQf"zmN[ 5F;=6Ȧ—X8;H|oMa_&L3妝ͼLjںnzTI \1ơQ=v2H.@2z\GJffWLfdnnm[,JilvoO B0N(F62".*wQuP6o%vu/^;t6,r|3Q kX .3|%)SnX>eϸJH767JkB7WuI^N)fTm$E&EtMC`LȊ1swӚ15bf%-EvoV&(vrs]JXض~&>|qnN^ ~)˗Q<ǽ3{ XǗݎan}Kl=H?jG}FIZe~D-e4r/$ό,wrsBB%S2i"VN$6iv G/4-1ov| sEZ3F #RI O@`0H4Zńy¬w0ie<\a+:d)H$6hߐGM">b:`T$K8 E!t4IcՈB2*8]([8~wG5N{mdzނo9g5zzR#qy?Eb2z< 2.aņ;hJ=\+Ad`M!F $lB>(H̤6K\0Lns5<\qv6A΅mDzp:uAR+q Qƍ2k JɄ)L$dĽ8Ð2ia%VLkJf=Am3Zʓ9Fqdg"a=0d&olB{??-s?wܽtt;}wWN>M3!4Us q٧hO_uAoGz-1N͕6N]/JI eq.: -w'??yo}bxzy;DbM``4@Wַ5;o-Ux9]G&۱XwRqW'VD`98!NH VW,:}bH4Yi`jVez }[#γ%h$ɢ@(Χyj*50 \z`<#?$4 dYa8WR~ )( 6 UXQr >>,ZelsV%7`'blek$(XI!/Y)'m R\ >'=fV`` ef د^(ug<0"W0&, 1n9Ku`nYha!+͇)\GX+~Ku.+z_(!EqؿqNX$ p.X wu1OǶd+k%Kh4[# <3bX1Z:9,Mq*f44sv!sCI-)cfK493fBEG\jW?1H126ɴ؎-.~]Q 3L߲Nk4`OP3שŦ"NnwgbtfD5Ev ZZ2^Z$H zj{ O̜/x: !njlmX 'VݱhԠ 8$.ب=$+bU|D %Gš4d UCs$x ^_7O] Wæ{,ƤI ݮBH#gH6:X ț9Nj< 汋P BS60aVTL(RA MAY~׻2C E /<62̌p!9)-Gz SϵI^Mxl%@&-h>D,$l^o=ťN#:RRdSͧj>TiO۩泖,o:nURF܁k5nM>ME:Og'mLf+Ǜޞy G5B6C .jɠwNɐ =Dj-U{j2@*t/ tĹKKve|Ixcqhh|ů{ݖt&A{J 9l0lT)Bd ̘TўX*٬1ْs@vMƀX95G#a9GqiGqi>4ku.'_7(1Á A\aB>x0hFp W'S=)NՂ(2:x$5Ml(h0q?I2o웻.> |;OOirqHpK3~|Q I IOǠ1iPn] Hc$Q'@q^fy9ڌk >>|s4bq7D6n]T'ԣ 5|M8g@tx _C(o0@%`^%O~xA>9 |gD=OYY2Gk͐/fMc E~==X溁ywM6Rsd !kRٌ71y&[PH+PzjhGS' BޜjjpZ %lKo&қlz] _Kl:)Q4L&Qgωߑ_8vCʻ+kߘs/˾! Moɻ4%bt|DUF7of/W4VU`#)WDbdTWcF쟼3ޏ`e}p./q"Ox~bTTO `tS9/;7F>Cwnk Q#)yp%F-m rǎB2 7Mg.fKU/Zypd&۹0$wܥ7?\Dtz+0<]lRq{1mu`mvmVg(/ ĭ)Ξd4@= D|(@ΦxTک$$ a]6_Kh 7Z%pE>hFtahe# F:s5~S;F3Bohm}Y>:3:6ٷϹڹR-ӎ7?]1h £j(@?b"{LЯQ`_gsy&>w:I!7h،ø\n5DFf4h4xn%wN: 9$,.QJ7GTp8^xe7?\;g$& 6x(mi!&֎,M{P-5]ሩũUU ҘEʣGciқ׎[[vcu&kD"*th ko+6*𳤅D 橏D<%ת+-@6rh F f$ f]9L9z{F d4 k$>Ki- o~cT Gqyzۭx;becf$ OuS%K43Q6kHaH԰ )(qWҨ})}a2kͯ\,0Φ+29 o@cAՙZpwǭAҊ~9*y&hF|V/~uwJX.kr} Om<ኟ\_=kޗ {}Hgv͕~ [s#W%kɉY}o;~I:_=?ug^__ze׏l휢zï"‘"{(b?õ,@ eSa?)5_RوL= nꃕ$|GP(U֝υRtR\bR@nBgkm+k e'xVj%V*6by#l?te>dϏ>D$NcW|=䁅{s,:k9J̇scϲU*Q雗Vͪ&x)ݭPY1r[/jʸ!=lTئ2?~㞯$ BDwP[fq@s '_'A2g%3 L*b=Bdk=W&R`s/+!c_B MFKl;U2َ,Q8:D< _<%:UG1z)0Gز^3tgt(z7:jcRu@!H]p،mO׻okxrX@t0gzl`PPH$;U<-0C7ⴠs@ȩ fNF/YrlemʔC唚ffM`pm4vFc>^_2I7:%礧ZZ Vl|Ô1#$\d O[*ULNFzV{%ԘZ1KOjEt4AMbB#S*g.FG)pL&A\; {Ub\ j1=d6O"OWM)]) e5">i1H܂cmVWm?Do.!uRҐM'j{ڱ&%"骍])G`+ Fh-4Rue F QlOʡUSȉC!ܲ5JGu:̀5{]"GC~\m:h [ݗRQG}xVս߾EGmٷԵFxlƘ3u`6j.̓@֦1ހfAŎXk=%UWlvѲ1QƤI:DB\|EQwWCxh-ߛ5F9xtVoX&6Aeч^G_0ҤP$fw$@cgW{A9c^$X.a{+y+뉕{v MIȍLJ?HrwPA yQ)5Do.C8;C8;md[]dmGmYD]^r-';6jB)f C 6s6SC0/SwŘE2r<Jr3#R(8ZVaIgԣ#9&ۛBHKΧY'u*yCq& !: 0)HoKOG )4uHѥEBDi"蚼<@4JNH(}Y|NF_sЄU13(YƘH=nlY%OVC'S=7HةG|SG1N`L%)yM.2e^/]L"uV1kCA&; )B1%PsOnFksSwm)9so>}U#[43hs1kZfeVH|EYP Y)Uޕ(AV" -nyf(ۍ@~FglsT 6=k0!UMER<2%"ݐ8/=5NZ7,=(mV9(OVeJE%#2nIrlcF16l #٥vH6xŒovm!^q5cs\͘j6κ]8tѯ ~RٝmT@GMU3h-jPb" JXo]rJM%SfGdIV0uT4SDP=u оӗ"YRwt'4CߺY,?8GG6ed_tBj%]eQt̸mqʽ3`YIC69җ" {MpT뱢My7+)X+4ƣuUo`$ oUQK)>, K Gҙ&0W.9ϟ}QUXpVfOdSSʱW,κTL cQ 0R}NzFVTQvel`,mѹ:ݕ3Mek p$(!; #}7ܝ|>1fqǯlHZCx|<9o!EkFC]0;#U[ǁu$G%bSHJ؊kQN9ǀṭpG:̉hfH{U!}TwGnWzjS\)%q-:{#FS {~Ͼٯo<( 5Oّ@2TrHf7ha :ñ_ GK7 to& \I9L@I.xQ7ܤtQcҬ>Y4w9TzKCWX LD v6GIwMB_Z :Zn KG-zZ@cboܬԹi_Dh ^t|i$$ nO#L%9Lq\SLG10d8BKnvRkH ǵsqo*Wy]^ubkWUPv܍$wV–';OFSSdnv*GDAccQ0nBn^n$wdcU5Zesn;Fn;h8@Fg)va(^J=(7Äv ٠[fU*dzFSUMM7Ͼō ~Lh7n6:?h&Ym0苮nvwo hA \B2a8Onv 0v?:=񸵎6G;VhBMC^3Jh7f8]-dV|VhB FVbˠMo nvSg(nX9YvbN~pY%"Uż[Rx|т=eѵdv /W dmVӱ풵o_;pȉpcʯ~L~eEv; ~Ev-jW>&V_3ؕ_6sN `5slU^!jW-FCnLg1uk ] 6}❕Yi>]*''.¹WTBw}>}yGIo;I_U:)s5ryu8Sj?Rސz Ы1ׂoE a!1B fa)6_gߨ(XUss!1[$#K62Ćw5DDhr\)~}7g4WU{JGVǵxOjуi^ {]^w8F6qT;tM5Vѓ^$EOxPF nj.AӟpK6&FA4UعIb^"aFf}/^X픬6'`Q_ ޮ~MMM}!? q 9rE1Bw ƍ9Ё̚PQ5ŽޢI}6HMA]dt&NeI͠CuKBbg{to:FVz ⒂HuZأ T ݷEQZr{+Ӷѧ/pc^DhMjl!Bb`0Mu8Y\PW1Dٷ"pV=z.(,ڿm|qUx>&D>\0oͻ(r7^vs0%}y K[6g!*AITJUMnZ. m_$C>Ԑ]nD$'W.i]{Ny]WF>N ӳ7U ʬsO{s #e]^1Y%jr]Xܷ˵'F:D=.-p[Ƹ4r͓QQ1dso䄎)Ha)Lk܏[U`NUdtA?OzQBV=Ooa$=`pP7rEKqiJˢG+{!uМy.o9#*MzP&uJ&H!JfIKq%z .34a~s1h@RHcA*e,WTxdME+1\-,88$Kw5%HL ɼ)sؒqoKe]& ˧"ATjZ+P2,6U*ŵ .bX3q^D\L0 SCf:`ilũ()ػ8rWs I @GacǍعMOcV[~uK%)UUΪ,=,em( G67"nhLp_z?~g /svy͘aƂ$+싹b3|1œ^jl\1j Z 5@=W높#WK*'t[M qϜ4E19`,W&蚒#sg[LE8LtoWX_$#5T |\@Vluʱ)ȿP,!IqL& Ĭ".&-4"O1AKC/,,Fza-t֏V~Ḏ7#)|WKPܓ c|Aq^v R;V_b^^L{CłM8j(`t@͘g dG8Jcb&Z>Rr{N=2@^[&n-[y>гw֎d Y<]Srɱg2QW%iRJry2DzY: S::xYYQRۨS\m6.V7AE)T&EuP- %{t#p{-Eeb,z=(<˜f=ηY,dս78RE}%o2.7WmY¸:VbN t15A]Ks*:dH)! cBBF6a chsb࠻7~:;'\q%u}7_-ٖJS\ 2(zz80#C6wLPrd92rY8{'vM3) ?rwfp=ͽzeS/#iWiٷTu%UmΛvK*d[r f)8Vw'1bO3 i~Wz̀g^kv䡅z2"sp-G/f݉EiJ" C2+ ߾!}WĢDKb1:d8ODH{KYgjżiL)Ξ"^t aJܷ̉"?G=]Ƞ݋x3 ّhX1mZ ?]o963- sWю%k~o ͘sܼ"CuMɱ-nk2hmQ -|vmVhT+5+kP ɩ{ɓe˵QlPU'' @k@=(//hi7=!`Jw=K7S_zü^~”7K| 8l +oNp X>xu=+6 j,,*6 %G?YEc8b-znK>ANkYo#[e_v͜> ylU.+ԌMz)5L%Rƒ r\Ֆ(,S!,AuNlӢ5O.W)'v36U4P`\mSSgTO7u_ aO@^fcJ1e%+>Wa4k«N-]5=ѡȤfmBSlE1CԬ-ϳx]p>9ߓf6Ņ cؓ`aBzx0H_O&LbMoG:NwJ tq~gkB7₿l%(D_QYqXT!dh贔hT)I"Y*ԢڡOy*{cI[#!}*׳ [)!! 8ˇ?6VNYUg \bgi>bHj9lI6ޗLƆ|Jb0 :ֈсAVun)9r=oku?y@T(`+hub x\旸lT7ݗ1v_3A3 ʮޗ?wCBa`ɚߢʭߐ3vZP˞nNx9,PSu8c˿7Zꯋ?ڞl)4ߛ pΎ|rs;="+w췫Lۏ$~؊1lK^]Ykh s!8 ^_?j䨀]' ykCgT#/֣76l S>չ~oam]=+i'RNcﲧs'>_>}k̺ە#2!\䃨﨧Vc咷T?o# :, "Nf̳17b2glRr;uDC=ki$ [{=\"rdmDWT5xi,1x]do{9Zn, {- ͕%|pU{f8 Fj_~͘gXq [!j] zjV֡cs+q-q<<9Fj&Y@.3̘|7K2Cy X鰘;E_ v~#>hisbnқPz 2JP}3f̳1sjޅA+ל%רgQp<(GX9kP@X9w5>F4VlX93"8xMX9#8|*}V[F;ì( t \=// ޭ )*^`z`LoZ{% 2UXu`sODa:F%Qwʼ֎y }yxt3fPԅn7Pul~69e ]x]c&2W_gr5w%|ٹxr/l .;V35~m0I Zp/6*^=62N R6ͻؼj[nj_zc^ cܖyY/b=}NO9 ,q*b.hEsb΢ [ݨ|ovv6cEAг0&zK}b {" 9Z)ُXA#\ 4ΰffy`Va0Qt=6{dUu%.Ǥ\>d5[┞&xh&- ZjE댔 K.ڢNGj߈L/T *j]n΋vQ&wʘ5#9ԓ@kDḼhA@d,_Psوϡ"9'1wh flQ18r "n[,B⓶c!jiWDBweKL2=C}PKbI~> |j¦]5"7Eۓsh .::n6)EM$e٬o< Q6*2g"(ll.1cZTV8:}5$?8l^' ZD3.ξ}Q 8q*14ckE!Sđ*r& 0WS+FaB[kD-Q0]uV_I`NaAtބ{m2MQ'؛ٙ`?L␭l Jz3kz}/{=lhRs},k 6z_*}Gæ\bgi\%Vluʱ(P$Xn .Acꍂ#gCH, T`QJ) H\u”/%j}w@1 «堜>S1P7]} AEPE*-0ѓ<Yo@¾a0n4u(E([:Ց݇В }x)KE]FHqɿ$Ϭ%j kֿgh?}!& yU{h~@>V.4M D1HPN82'>i.^{|P5;aђ:o36'lMuh[͏iĥ*ɓ $/UXؑdwYwZ=`m1b$o[ߖy 5a-޾=Ϡw'D!+2- @D%xC8;|0h Wh7]czӦs񯳴 z[ml0@ڍ7<#EՂXZ? ߎTƃկWE$JcMmAnM1,5GFot~*x.xcJOb<4h/ ZƣP L?}Ƌ+s cY_û~0XWkYeXҧr*v2 OիcYWWBI8~5  -FȬS;~vٻ6$UmF$a#r.=#Ӧ&)Gv~CC eN`D5_Uuy!oEmz Xn@V =qܓXgMX^:*U0$eV꛾5xȶw^{ f%VxĪEc:GaԎ3#`l4m#}89}Rfbr~bE~LoA+sơhψfO0׏w&5rYJ3.b5pʱn4d|25*5#WOۼrDYRMDkga8V|w5WNKDbRDtO,A&+I: c)븺@q6K{EK(6jQ!r͢)w_,+0绞RP+%K]0] $/9]z΄7^vr7f`=ljK`-(wBXF@ uH niF5 {8Q#`#5I(6~T^ ,/M ,4ed<_k1#i_D5hTGI޶ ؾ%א x=JոO½$X: )Y]Ri 1XY~{^1Ż;T<6F}0QxYÃ4WTWV6yk`@>inҮFN1$TY<qK9Nk6rˠ@wo. nUE`&_@r. J%M/ "t%(֌":] >iطLp^ Q-h)%o:]'!MAf8^f莻۫u VetuDߊbdlyp *} 9+80zPg=q {ޫL\4+\2_- 7|s#U'7? ٫t~ʼwjGG,%o0O^e 7Huc3TUjPYcMZO^ijC}.Z#"?tq pp㟧G8qx9$t x+rr]fCe^4fq Y}ߟxKBӿ!L&m2ts1qA]$ /X̿F]-`bEQICִK2b` n-F+# ꍝ9>&0V[nV+aReVh`^g)oEF5I <:_+Fa- ,P5PIiN{WgN$]Q%kW4X`AN,@ӌdL#4x ^k\"YͶ(ZZ˨ +t)F@64g&%NyK3c0Isc֧_Zz)˃ jXJU[%a>̻^qIy& Q^(`ZW^ ?x|R9l"L@(Lc{3DEؓ(/5OhAȪf V}" `5bũe±S?ڛAJ{/@KIY;ܦ!) ~?/XunC8\¯VkQ7shlo湆iRXv<00 |mir ^0Ʌ23׃Ovd'rp-* Rij;z[/T`&2_MpJ<)87tNP3{;!vHt9w␋'s7amuu9f~XCn'~߿͛׋'_0nN1t!w~~譳VRd_=esVzi<M°~]TXUu{xLlj)i/W|M N9< $UV:Mʃń~Sp<êܳ==9 כ;k ǽUb?NO' Of_]=Z Fk+D}0W;gj@*}hx V8lR}yhqwwKO aS\Za?2 u=NFzg2:.ū;PRo7{Bn 2,_gן~ߧ7օy,g opats::k6U2\[S}pɑӫ:m]RL,GɤF{Ь|EWۨkB0+2eJ]+)κ25ikK{s#1_*X$qQ9Wi?B~1䲽V!E((MVIw+h{UE`4(޻Ϥ}L`6 qd([%3{Z}E71nbGNJIna(QpQ^1 _<+z6vzşrqE\>[ wF+$c̾tGiDP-Z*F t-rOwb4CtI5]⣇:a6_w`"0qdS9o֩Gw㫖w%J˩nSUŖ@n i^}^uiJ*jcI.ELctJt8]['>Rb*VJkO@=ASB%ek'D֘6/:U%,{ⶊ~tԚ~TAl:+zU}ۭl&^" V ) L(=]lָqi&5 ( ށ+])w[lQ*MXjFDBu-!z"e%áw%᪸r Mg ŧ:ۛQӝX>PS&jǯlhA܉T̄D꣸\9m+"/ WTܞr_^g3;>yԪ[`kx0T/CWBv׿y`ttqu v.WQL֔FiZTm nRMf=p =kgk! |./|x![{A{OA?7Ϟ5MIMC_|mæ-o߬,@KYg{҈L&Y2gdZRW+~•Dymh)&Y\cΎ1l0NP"E&4ix˜5c܇#͏4ع $7]i$[DFIf1&oRM GED%V}U,[ p( ,땺Ǎ [FlZ(sSrwxuQtMMY5e>9yHimOz6INKW { Z,JE[v ͗uZ*q(o,ֈ0/ʯm-6}w:xj_!n)%~GY %(eӜ Qz<0DfC8l9/ofAs!HW#p46cpX\`yu)L-o ʦf]Ma;ԧC9V䌱f0>$xыڼ֌UldLtܧp<}yL:$vf:AWU(;KS`Hox_I}y.}į &c@zy"M Cc8M O&Ն:> U1< V+۴@v`8&7xΰokL\mgaSk-xp!oOd.'Dx^agɻui?0hPG(_aLg֍? Uy(;s>'ƽD߿ ̸, M3%l$q)pIgI*2/5R)8\Yfjrw:N&*}PbPwA8t;p`ث1w0ٻ(= @}2Fgz8_! fهM@b,S:8Yo5IC!%K tUq){)8lXoCdD!е x rviH<&| u‡(~Wq|MM޿jBLBT7ڨX~2!Ɨy-W}9W7ɘ&O7ӄU/oOijO" %!j.R.G9}4wbg *bc|< "E#^SJ> ?olOp|61/B=6Ef.9&ȁv 7\P8xSr LJ*O:J8J?_|Jj $y[;݂0̙2,' 09Ora~tWqe?WW ca_?O׎9GD;H"Nr8x~tԲUJ Ǥ.;SV1<񰹽p99%ɐ BjMh[既`Q-;D%VqubҙDu :=C :ef5kt&(scZ Eqؐ3= 0~Ry,_%saS.&/݌<2탹\2x)BJcoD,cD6B`]6_CZ#sV@[qsoAxQ*.7X;[-b1PMx'1PRvxڰeQnڛ&'7c&r+j#HLDSYN[!)U%RMSY8of8)CjmJWݭ0B21\v'm/I.I\g91*xS*p&@[ EDD0Y$(;\'M8{0r3cp}:gىTtdC&~vw/LY]l ?3ӟA.nx޼߾cr/B+%2-AF8T ?ApEGc s=z#4C }2X[VYa:೴z1Mm:ӫ{K+Ur)]cGֺT˒NC#}(( uCSpDF 騐xXBnU]bIJ:SPkYᓼ)Xh[xAK0)LdA4\G8 ׅ4J,t 1dtw,a|<=AXJ.H?qIKZuEHF$b}5^#VϬAϜSC>${2Sr{_r𷖺*4Q!PKBa3,ـIcuF=10Ia Ǚ46(FWf?mՋ˳Yf7j^vZgˢNǟφ>-g_s@k2SqX/QPsE_[7Q3:~-wzԅ~6G_.`ît@)7 kvo̽%$DNxH'F$ xRsǥ N$eެ_Ƴ,\k[L FP$^<ҝd&_n3xaTj\@94\nFy0Vu`&J O2XfFp*w`V^ݺj0PRki. #gb Fh(0T ڨDR6QE!I QGPhi"%y`iθw45]V .(4C-RJ8>|4$nr=eFdŸZ*\厕~ 7LS;5<ٳ<݀Y+|'*OX/5h@N)qIfD "H(ud݂q3)bP)M1MdךF"HI20r6 Ω("I$ -@EMNp17@`VFB;R[ 5TRO$Nꢣ(Gs7锒Z;Ƽ(0W4:CM7zo4t)R:)*5DY5(*=7 nFlNoMj1SO>Wh'TNB!vlSs..ն'F8%:j՚yaoV cgth>F|{1>o#){! :zKowg]v`!gUutI`L}r ձm+5>za/)_]WZP䎍d4dg蓻=lv鑍CZmvrٵr]eR'1>9>4珨:gOKw/{L#_pȯ8:*[Fu:~{y`h|#J416_%{͞ko$ZʎܝBD @.y{%noM f|0oR ,gfbmvspWɊ*-] (G;.zyYX@DEeI(6@d@/7aDF!;8bYe MS'RR.#L&$+F"rj\{R2KdB}ƒj8Nt%,b8W R!6a`4&=܋oW+9.+k]P@&lb#n5,^$xUWhA\nnl(fX#\&-ubQ,'䈳_+Î 4+*_+( u)=BmmWslwۙOW:^\(Зt}J:"..Bg&’԰ctOQ7Cӎ==nʟZcZ.d!U `>YP1{XZRQr(i\6V]{ƫގK^g8Em\q&(iBDm2<"̰SP]%µIQ)p\|P(gK/ L5D3vޥ@0{\%0JX-)fT3Adcr"} &xs)^>'=Ka?\3E .@DtKn<@ *#WQ*26cg*ynwyr(,_b9jopnMJ=dOJ&DVd9Hd* D /+RKh/~O9o)/1]JŔvmXwG̐V dTCf66Vƒhݜm:VjCp]IkAY Q;S(L2 x7usťd׃MŔ@B@ytViT7w2xX3TTkT̩ɘ@ƨl]|͢51DiVU΃{)ǨJ8%C*x[-VwT9r|oK]ƀU~y/e8r"mw2w>}.[tq5J(n)=W_epiρZ rQP*\l9ɶةKVFu<6c65o!&ԃ>րR $MWJX}02֯#/ȏ0)" X|}D$C Rz ZNC#hC/q̇eϗ'r&\EjMNm ʝ<M){F0*<q<.=l|ٔz:Im3o8Ss VP嘾AX+0.d.F8*DFjVBn~UDl*k{05Ҵ #R0xG2qBh#_4%jVR_zR_wLеМ1ai;dªj*?8ڏ9,L bu+ۯ-SU7o+,Q F)$v`_?H;0c?bbo$&pF~dD$n&!3Jn`r=k48 .s-(Iw drRF=lK\CDXؖЦRȔoe[4?= FiYO_T<\ts$Q| 0cijkɏ߿p4t z{{Nt'k'핓7??wsp@?W[\-'8&,OM!g:+{qE{Wswa&٧wCH_'^ŧ+yk_ t:^0?84B7~N^` qqjWQ _&uxnKFӝgIou?Zhͫﯽ Xٰ?nl4ΌKXaf FzuĻzԗ~WaZ?mƎMo '@oGVMN'|q Y'سH0Ƹwɠ7̆jeen\ly}1mg6gޗ*uڵe6Z`(6`# /"?粵 X"`kTpCqf=\&k΁hn#b-Z8:jL7;NJ c'ضǏ=稯(sV⾵}n\S>ՠ"BJA3msUL[-XnUA*e-CHSH)jewIgqs5*b-Xv ObmYv uv-\,@|8HHL@!XU&)8moCEm]c4VKŨ 9T]U9B;lJH-C£ `&lSڶ( !Q1(m^1DSضE+n5ֶjˆ#yCwb"7_Y:k۾KڜfwejC._ $1,`f&$ԇLzD b0ќzCzZa̘" XVL2תh:uM,AxwvCH aB=tFX٤^8l'-f:&=gn4}tؒͮOFzӚjo7όMOOƽoQn\ƌ[SGZ z[zaQK}VDs>ZUV.|ڛOoߟ%a|i=ןoqaux}fldR]' L8gW!̮hn41Ӑd27`0* uoxt=No7Sho4Mu2Ba%S>W8G&azt>|+<'R~筊/xBC?LZ@=KFExF0X&[WdCo,gcY;F9UW׃u3 IY:Y;Ky*)h4K-kuL?1GN ; J1xÎ-#vqD#KB~kGȣyyL[AhQ  )2/#ýܱvG"t1Uh#4 9!0v!UqΎ)_6l##$odIHrݡ'` JX( 1*RSʢ?PH'$k7[gK`AcOu'D=Q i"(hy56l6t vl !#Jd5njQ$@"BbdnmB޳Z"~t1U`#08,o%^O B`NI B1pL5 nU,8m~} [LйbJrcif`' lٻ6dWz 6#"@ ۻv7 $WD$89?38HyCʴLĈl՗뒝Χx͙uq%;7H,$ 1Z3 0aaށ e Qm #Z1Qցa*8TB6DŴrͫ E 8oh߁.7U%B2$ǶTc2,f.uzNe)3w#T)ZY?:Ae&4R}L}'0s9} fLӴI刌)#9v4Xɬ)c$ Fs"I{Zˎw TpwJrzr?Nvx#yNvd ۲HPJw#ꂨ/i)4pr bNv,6ӊ=&v٭O5HRѨ[= vNfm@E*$R FSp0QN2FE4Zkx$h&is~=ߙ%Wecw4(ӯOU}7mm_Z+~12<}k.#djv W&b@<0%Z PKƃVɃ:-I[38w]߁z =^Khd H݀9" pjvqqAѨ#aECTؓ$WH$0eh!1vB7PO/~, *M;N^hA%q;yW?­9YN֫/WyzE(MJTeDpezAXoMf5ɎjQxkxz^,pa`___N/P~2Z]l?/z3|Ʒ1PcX&?iTP`4]JڹI\XD$ j´ 8* Y ͞&QqC@{Tm8J2wd/#{^6[]u8͹x~iĬ۔G"iPAS$Q=a5ӣ̚z}1ٺwJp8gNhV8eV~אuRp)#:yPЇHizpm ]HHU֭P=4[nRnTqo~hplQ'~Gvs{y%9)=?LgKۚ/ϲwCY)4Ngӫ3h7̭" ho(Hea:ߴs>dz0L'dfՍê ToaF>R*$56hgqCZCQOp`. \\"TVXpq[-\Esf0\S3<'TS9`K7)@qnJvjl4 {2 Q^{;+N^;1jOb "0|s}M%#R1)ծW͡ϋo+/-+>{0hejQ.QjTOOz܏A$F!izVmn$<(zʞ)=UJŕFONY- Pe7%j:݄NdԒ'Y2T\'ʁ(N&=!3sXkh<*F9m7A :UNjVT~ݽ@`sPp!}9^S="%MpYM9=ɅTz:ɲ{'l$R (:OE>u٧~}CMbʡȳ?ݗd5ZMUcM.NpEK*)<@D#-Q.šp>MN=_i)Ժ+3wӺ'z1_.C{>k c.e-f[&Wl{o~<GMfr܍or*g<`:8_g2u(FyaSo ~:Xmի1}Ǩgt;8>Ba:;iVliVtuM8˫txW@r=جTg|,3OI)cr2-bsK\VIpz|ˁ,rRnW~_'x/찢-J.S3~S9gȏ)uSzUlc8j~:rF.}=㎘`PQTřHu9NKJ?6HWqxRT# ;xygI1 wteQYQXm>7'+ 'g_ PHv0]泒+V״qs/e덑\(򩒰 .ޗd<PXߑ%%#EC.z'=[6nXVEIbQ8 L+Uq,N$Ym_^GP;On-?oXT]$dHVqB8Ww`wn N?~It2fv6]^΋STi4I-j;/E3t4;tJ5'UK2BjGCj rٴ>\RsI|.j6_@=] x=PI7KG9() Ħ;!ޫ!1yzO1"@@GliNBjk$*&fN8+4ғ i$K8Q0"_HF7cƆ:j]JN^Wx?wP%hl굷v?m1,y7(摱4{U1Ba`zmbrE!"}xqξwXfwK^Wŋs ~q*ޅ.fXmQm "$i/ǩte/:Oշ_.l`Ӎ=F]ɚXjgTɀ` Ƈ7y]Scp7W^gIFb cKqa>w8a;UT ٬f䭡^N85*I \XP$irQ\9ƿ|6q$翂%g'Ru?\y`A)K*,@q.vB$8;tOOwOOwhYȣ%t0<߭ ab-ϾNwnce/yʕ b;oY߻.{֟XJ p̉[ FD38i"M5DvT#Dv%ê g޺ *y"5~.qbi"ESJNP:MN 2TNW0LBfy$RMvWN$ZrS/= ' $ S:Qs{ :( L[DI!RHV,c(r] 8E`5l|6˕ tپa|c/5I&߭QgJoXbNCH`ǐ{E;$H\`U}?u_jjD^NXCyԙE>Bl|}'RrK/J3{=20XNpkߩ9Ɣ(/VMCSS N&-D\{ʉB ,)BHEV Q}!)HP)D#P | #ItsgIA٫TMG+*t[˦2tAGL2p+|ꁡ66@uVqb.P2M6w<90A$'  Vl$JD#&&5a4AhL'QSzjPhO=B t7+!Pf?b]7WNJ$M{gW:m[l#FJU2sۈ ̏-|Wp""FcC$ă~PPaHT$T9\ u A=GfZ,ZjeU._wo'vEja|[4[G[m䌋RV䯅𑡏t!O^#!e|3AH%hl'B\\t E<xN 3"&"$ Ɲm{n W'mҰ.k(hr6޴4!fS';7i/ѠZ(P8XyOAյKN4X=Weo42QBnL 6DQuOah\^޽FW Y{0No6ݥ["n oQJU(ʗsq:8|AܮfRUs⪟;\5ӷi U>&Vh8촜6lBWd9[ >0ҽN7f"7S3KMIw[UdALtM4џw]7!l-~82>H :ģ.,2EI4Lb,EZV>Jnw8NRqFd(07jƴ9;ʬ8[|jIeYL&ߢVv6./s}$6 [zyfJMf C;u>e/YR]E? Vt3Ņ׆F8 x)+.R; Ц+mn: O#PW-]b3\:cTzr {p6[O&փ2)Ceϴ֣>@y-ˎ~0 ÌwӼ]scO,ճj}xSsG?+5?kqҮpsq4:jqjbO$:Kݜ8lN%wfWp F4q~|u]Vuv,{hY+[C7Ř(PC{6Dzuks<ҌS>oB5zF߈`C1֮o߰+Qo-#Mhˀ9U_|/ܕQx [%ΌtRۚ`'d i3Z`Hb bhD쟜̀w淟m}8!5Wy88ǷNa@/An:nʘ~{D::_+hidg&yl-Ns")lT|OQ@Hy-ݯXIʺQ gqq7G|lg,NzQ׈?/ qxuS&H&|c|7|~YN:SxVIasxӷIx=ׂe5:H"q h$rkSN׊ &_?}3򣛐>aEWNt:=CJQss^ūA)p?_σyR.oC̾0  X^Rh*?1-3l'üّz/:q}+n|3i6B#l:(A*gWΘ@%KDt(JkldE/Y.}3YFu1,_?9+VaUH!+Ś 1 ]^bWJ${yoOOsC:" `߇Rl~wKHwU&0 {K$@玒R񓇎2\a˨ OQL&ffnDW>$c-JZw@؛%=" ڇ@FWS^ˤ&"3>,D!_:RAf,]#j=!Q1;XueH!^Z֯R5ߊY ԌQ;G8c4m1 yNnҨ  Y.<P{2%`̵ؔ!k͔)ks D.eXkC$h$B@Bݬo(.r.֞@{4d#~V>t^zϳR;cC;To`-bE1UE˫Gqk/m[26VJpY@y8\ B<@<|P`$Jn06yDhKxޘ(Z $33؃u и1>rJG (FHڔ/@1s?`+S`ƃ\drb4 ~p dCf*̝46i9X-iD$P([ ڕNqdX0~{iwF=wWQJp[F٥dk{\?ަ*"$&0qR@9 UXS`/׳TS).+ࠁ:8q@|T@IjB8k6k4փV9T&%#i61jM%?&+j>ΊWppuɊ!$0 |yQSNs̴[! `{:6 Q^;\BQXW0/PaX-ȎY=L*+o@@0 Q3?u#NC0 ؆`6)Jйy$VZGUᚍ>59PC[TZ&7!pO~Ɉ Ai[ }.c2 !9&wA[ʺ k/X˘#dz } ,4eH'ku#Be*(f3\Dlfإ3L7.%>G'򣅄"'K$3P ·E_@!PQ\y=8R-|h+.@TQư!"ic)AˆYԖZL-ܨi=ȸ,K>puA.OS3hA.@J+Pl :A1 T}X40;皠ȶXEޑpa+罵`t2 H b#ҧu9`Cjbﭼc&P,4>ǀL(=P ǡށTj6X} * QLj9BƂYR<#ܨ~:s@13"gs<+ cvb+E`A ]ɉRo|Xz4$4@趢3cI^PɏIcA-s@u@@ 3?:"doJÌξRNZ`4XMBݡnK̍L D0@0uJpuAΗhf~_INBa;L ?̢j4/ gF+heQFIhv/f cf0=VۉfQ\uH I z (9POQItp,*=~. ":YRb7T* Bt`p-WTi9eG01z*I86*tOGy[QWH-9at8P8B. Ygg3jVꄏs*4F>T b$0RDbtKX:@ 8 t]~FWh pNHO̭&L35cֆYXAkdǒ[':%ɖ@&s=% ܚ@$;-|ic]#rr/SRΗrmRz]幬nx}z_pBg_[{{grtmm_^?[-_~˃WU[loƿ^zo*]?zk/6>g]Yw8+#wք%w֦r8+Nؓu].Ł^lѴ47mhnzau1wwv ʤ@CzzcQs qeq \̴4$& ;Ɏ#=)q3wز  [))]\@Oqxiv6}XSM>sFH۪?xn{pox:mosp `k_X%ܢ]Pn6{-Ō(e\\=ē)HҵD8u 7nr&nLv/6Wll^l<آ7qy/6Ullb6=4dK.O).[Bw/^mGl^=ۦ?3ދ?HY~syӓ+?v< '|~&#Vں/>‘ n"¯^|HO}q/&&~(GA]ۡtyuL\;w%LOAW=D(.2 嬴ނ]?3yw^? ;H|nc7x}kՌOå1ddy߼=y{~/OZaT{C]z<_^}w$Oi&cIk8 !I9bs?\<텯R}|hMi 9FrS]_uWU=WJaq{]dwCmԳ7}nZHV6LWe!6rwCkc]*wH_16Ējg hf;P\h Ģ()RU%N{IMfU}^H+cUƆ ARR_ҲW26#lzelSe".f,~ɒ}D"iVL;NAbק c(jKTCꖉڡ92A2$;pVJ"wtl 7#8U4FӦ"-[.>PŨ<8wٹaх,\J+h*!BB00&'5ݤfqLy HiZ_^\,/__dͅJހF&q?hvaF/HΊP"b[=Yq(Wj֪Z-]f'"|U5m2傰$#]EC0&m*HGԱ8 mwyՋ@ڜ$N""Q)Ol6 hq.ADI m4/n_,6~̈́.?ɁֵcA|6(~V_?$ryn\]}W#Gos9vnX٧"I[Ś+1(NabbyNci}Ak?,euk?zpUŲ=w {| 44|AhJpi0)Fك @&k*,N/vz~m]lQnӬebF1yAze&ͩF;]ݝG51 *'t;:5phڒwOc7On;g^C3s8s6yty<0xwGXeu8~ mO6 N4[]>87{y)F>x6%i˺zjSMxGm3u'3s333B:L@CS3YYvƣfKn|SXfXn\^XϪ)^ Vu 넡`i̶$Cٱ iX#{ayV}nw)6ֺ<5:8޽rm^CGeqo:>Mo5Kdo<$ĪcRFzc-s1#aXzl1>n6fC+Iry/솃4LM7 QڒPL0e`Z^B6 Ǣm*쎄3,tVQN_um_<^:a:un 4}B祖;`{'Y'=Ui٤Z-{ ^Z'=ZKxIО7?vq#pe-j’z}.#FL FXs.)F˺TCfh>w`:LK l2TFߟޘO~H?)J=?2hv>lhڣٹD\Y=i֣zbhv*tc Ok>{7u:mfN{ @p6֖F cࡸfN㵒p=0tk'F{~@[b=޲},lh>{^a"2$ \ ~6yo 9yz };w@ŞYٰ/̆xaE<96%ld&lK;FFXsL 3Y'pޙif0=3=F3w5;"`)8f[}Q~@v<=މS cL? x{: 2}{RI pVĹ!x4}ĹAxջN̚IsC`5Z{ *ލxKw#ԲoCْoOxh۩zm*,l>[ gR=6NGR&Nx|=&n`\V{E; P>_.57|8=4W7w^oo]^U?-3Ƙme}˲}],_?_"~\WvǮyo?vW-L_{; 0X7N̑#Ijjߢ1>:~qW}xq10)>HGԱ8P_^$HR`& .̆z`'6eKԸVP 2 1UqNY}L:6*eԞY{^q51.~dP$v[Ȱ_a (TRcJ]ZwOEI:@D)vYE\ &Z]ڜ+xA1̉sᨡw[12-% J C:G-\w(q# RI&ȗER"|ĤLs4RB)pcPS`R@W5K85x.-Bp3Gcc5y˰)aR9oM0 N ,ww8[EF|OT,%t1(*8Ო,*/؎j _KtL+ AX*]sicnUqP`Tox8T0V,}f3"sb 0B[0( mǵY p!s`, Gߒ),!HJi\8EV ~ :*`B ~P(y6IV>0wJ1?rN"2U1"$lR`#Rwsnr XL +ّw^sV\@tCPD|QеrtoAgZeL<=\0?޶ߵT!١4fnJE.1 ,IAv%SNihm'c8@2p28 x|-,FeV`*-ۃ [I1c9 k[9DBvarLa@ .nźda$#hsRbu%-j7lFWP\p$DY-&lJDgvI#Tj]X$dAP`EeF^aJ*UjV [1* x ū\a*J7 lnJ[$H4}YR6A(@3/hwTw =XXieʸu–GM,N4`5pX dRayΘ y!Eg1LWknb4-/#|ˢ׶902(4[࢒ŊԘ2:ub!4n)oc\]gz8_!E(b؇8;1dŁQ]U-sM /5"(n%Gd7Թ~S{;v rA iȍׂ%A`xtmb:I f?d^Z$=͆02SDS!7áyUBjt8€ (`z:6N&@ըiThuJ d[1d L)7bZ2ʐi.LDPfbf~$'ȴ`_2CCȠuGO$Y7F$QRoXU؝$Q0 ?H΃򥀒No&Y[~2.ZB[ NH&'XKHp`ac֩v@h4mDNThoX K/m3CG/pߴQasl&Đ(^"Fr@.Y@4(tBcPODyFb]0wz@ L3$ QZ1jRyX٨a^!|GA+ CQTH!yrdpQmLۤFkF{hyX 0ra#\6|KiGaA$YyTJFV0 9i_$`.}C6$c>ђhxk1׊s͈،Z*z@6R Hbڂ:&@8IHSR)Op QʭXizD?L71}4w&Xw߾cͲQٕ޻֟ʳr7>0͗M߾}8+ZӴ좳]eZYwOSrQWx>{M:qm 2V.Vg]`Ĭ]U=Ĭ-̲d2K uz+ˆ1P0,Vٵ`/z1LKNv侽G\ȝD' qo޽gn"yXPsQCPM Bev_f%&^fj@ħMC6iPD`6[ԒiE[-jI0qlqud)U%AO~lQCYD{jܦBx7qz QRv_z|wO_Ũ\ZKb~r*s!ZqT.HߤU[沂YYIx|m)%ؾ?PY BXj9GUQqiu^OƐ7P޿Lz{_^ByUDz>4g /S,Y2c jNG7`Oor9zZ NeNkF=s3|H "V~I Fq;Azː#HdX(̊gfb_OA8{k2!^B=|L?NJO*{.=rMW⾯xx_мEwsLӴ$|v?RΩJDS.q !Lfe+vSQ-g,ƹ&{}vǎ]C>%J,Iʨ'RɢTj)d6_t>3y io1BaO߽Z8|lP}sl_<,)N"'|9ԓ}۟} jc8,'1 d ?(f^]ܸ>,Vfܦwmcę\spv/g2j2TL`i>@/SPMTl0͎M ԝ +1hų`&"N`ʉ;τ}f?C|MY>Pv[_([УT|=ĝpaTRd$+q걊2H@x5Kjd*ug \O$ :5{4 Iirmg= OGFu*lfN4{X);5{t͝5 o$gm^=U 8$'C֚r̎0"M*mvEYF*p[#ѓrZ{- WZ1fp*9 YFTHc# 9dRU_oH&zE4y<ҢlSMuCOUvue`UZ0tUU:@v:H`-y2SU yNNu6Pu!1m5ZPFp@r>S]&zw=K͈b.ׁ͑ȉ ,4XFM9qPHQ1Jxm=09%:;Ph$Tcdd8%IV%J6֖|JGSIf5å!)})J2.d(t=C1^FQ6ި;tv%G\E'H5b}깛 ;:'a0ri14M_ga4j_ٍ}L_6"Gc(q{-̩z%bZ`y$ZI);}p;4̓LB;cxD(p{Rό!+u],'&ٷ!t4Az3]SoE ,hq[&֧\r8c,g/8_}-I#g/?743cߊ=tn)ս7Dޯd2菁;IL1gKTPT+\dpnQ(IZ?+WXk*vNxdzJΆMdl0U]emof*G~C'QBtI̬rIaD_]GEȖ ._ƓFW鐞[_Mч'ԟ.,^ՈF?l:)'qvo}wy}j˿$^CmS>Xs>-OB - :m ds;-ʑjKgIjV%YBedw: qz2[q~3r qXw+}Y'T%?Mϯ8~kt7/toÃIOїr)XcǍR]yht:[6^? ?p+?LW7uA+jE̴2b^:rAWyI JۖՁR]]qt +nZZڎzrȖ=ܖSKRm ` |$JeR-7Dɵ'zJ%Ú4㒮rCwut|IKuI?Z0# +e^l׳nyIEXw=cG; ȔRt|aW{B('vS?JuW^B;{@y_^tlIS*}.0ty_$]tdI eWb/ęd<8<{S}G0ہ["mp荻HΣ#ضۆtUӐ&ܶ*i>=kiDamjmTmѴ*hC۞UVU˪gml9+e]f֘rXeEX20:iSØV=yHJ#BBK,hGIn Pgx-d;)32-+OAY! 'YO?RL5S,.wSL' EcULg ?lty.vhz1_ =_W(m\|8 Ǘe%a@ByHqFP7w}K;;􎦣ωd;zQE8:҄uozGZVMlԴW_2sٯ=93}as5O^9`E[ֳqos $G#%`|WȤJ-Sx=OEf|2-JepTD~&g ͕4P;!;`sw*H+Єě[F4ST@Wb1h5;j˜dzii[z&@U{d"$Q !s Tu{OBBqw 2&*LkTK4+\Q ) tl8FXO+ٗsyB ${%GG7D@k4USx9 Ǿ8 tEaQ&9vh]@$ЖC\#d?fǣ_۟^dg>T\KB+-miK ?w'FJm^S6{t_'rAd,WjB0yWz5-iUc2i qʎns5WTM;z]Ff)<Ź7HtuN>2yTIf0^|dpN.J/,7hW^"wo"{.<13VYk:jԜ5jnAe8ȣ@V \"׸O!WdLnD5Hfy<`vqFgb-ז jj5iŕD 1t:~ݧi޴tú p$\E v'~:j2O$ăٹȮw$ vW# roTZt- 'pS>)0*`4K6`^lsӨOJiK<[2h,Jc5pFM3^"Sv#/x۷xIFX4/X&m {I6kVt :%1å="+os%@wڭn=%;40VK%PRaB%(XP?0>N`))`lӹA%yyKEoЄь})0R$SUp0lgZF;.b6{gsSS=[ YH'QUjЗ'HKIĨo%B:·$6XGO`ߋN Wݑ#DɆ1wt\"b[`qceք2;nҷ"qh:zWʓp:z?aajG1B⦙A,]*Q$6rT57y|SBC2 K?OGeoۣaݟ~|f].z+XGJ痕S5@\~#L0AU>6o,fKqѠ›|~(Vyi]t eo 5`tz4{cJ]a^-g;"HkVb{9ؘ3Ea+`f0RL"L`s>~n+w{&AzA2,4q9 ? D\b !% ,v`>>\]B` ֬ǁxDs,g= D-S-XsFg=DPiz6va=.YYozza*1R"R((g>Pf> xEö0,`d6r'`h ϢU{|Nȅ$W']T<-YUWqjTXF[>&`'>Aw42ۉߴg'TM0{zDO Gg a^ԤLP˜#L\gZ#AQ} 3lKUBb`QZ6f3ʛOzVw`joOR*yCn%R1V"5GO 3󱱄0w|tL` 2-|R p) KI Dt`jo^q$ʸ׏W|V0qougRJ]ժI_HӮ֮㏧4.jwsz!7]:Ɂ dit>E 6yjVH`1)-0U~>6I),IP[ZD@{.}wIoڅ)kUXZ{zX>o${vA7^2L[OfInvPrJv4\erHTL<=nO*12ws^R{PU6Sꍶ11}m$c{ єzHd:Rr;mETS=$eo5EէU{1=2ܛ,fv2|otvG9{tMO>wGk=GG0R=pcCe]Rucb:k'ZYK<c SGw#71C?Lp>,GD/lBYŒc0TL&" |;,_] WlemV턼Шg[x<|z@^Q";"cWTsШiG~$w(؛>9>p"^l]zb* O F1~3]dGOtsL{445/o5а;W6;/,rijA@Lgmd)=Y\)=Z,>펒f2w nD$K'ȫ9XZud`ঙiB4rp^: k3=R3C}(u"; cFv genKUW EHMDf6-/yv;޹Ph ImBչH<"x(iVd.g<Q\' YM&y!$mEZ((4vmIEtvjt#I-B׍$@֫> PYu,&e `>4{TC5 uF3jb\JlQ -asNg7?-q0G(dT&$%G?%e (͓b?(FSWT{)v)%:|xvMYvK,ٽ&Jh:N.#;2eO=y W7ri1_ٰG`Q^yClkLFƄ^y!1R]u} 4נ䄆:HI8PX`"L4dBg[):!zJ`GEu= >+Jd)Y) < (mexzqKʭ܎l; uq,OZd"T!@^6Jʇ,@ւxnfΛ-qJ`(^Bq1!l*$EH&.'iT_ qsД &h}9@$y#Sѣ,B.@Dsr@Bc:ޤxy5%o}WaM-ڃozmf}7a&98G%IgjP;L5e,6d( nF&ӁN$@>ǽoF,JDOdqX%Ȋvl]G6m,g4@O5Urb~XC*\ooԡS C!ķnM>qZ2m@e4s.C%bLi9`~+y*g lϻË> wato /[l<T~!Hz~ , (%dAO.((D|) B=Ո9iN)SQjx)s c,Ԟ/}˅ʥZ <3L2gr44D4%X)|g7ցXR^ÿy8[Ȩ)&yBU+ǒP\POGV~#k8ŕr8'At [Z{(v2L^D",e$ \念~J""ombodq[mS9#JZ܀™kYT9 qL0 JPpQ/F\ZQ7 A1!螧ھ՚ ztszBfI*{D릹/6>o! NK/0Q[%i2KobJ~icVz儝UFt[GYjR1$٤97KbO?JFAYɸ/fv63imoԒ4B7KP׆=%ToՊ&BinLvJUDyin`G>: ITDl2(A02vs-"|x Ą0_/'YԻD|NHw %F7f hv y%%1ttHz K=s8o #B/Q-I߂`G~ x.}SR.݉W\rW]C{0t&M3DHd⾉HFUZ$4o$=PZ}$&+oGr흎&6ZiI27/Wd}QMz7]BRW.=Qקxzw)pr#ոd}ێ~ 1p~ZzVg"VWZai>->[ /$,c|{$yg)XG,e,X %=5ȯb ;)%4Զ%HKƫn|bʀ*-֜ -^LC;V pF4m=ˤ?Vf4*:?GH+<Q\[kr|!0d<_??>nF?a̋?/Ҁ|X<3x{SeRm(%7p@ ԧȦS8Z׏FYq,XKQ2dxϤǬ. *#2AQ6IHq3^m@1jU,]_X2zk 4! i҄+QJiӄItQ˿uG@#E)xfb:i%a`Z]~0#J3 Q]&z2{2״i閠1X3nfc͐;|(d8?(P :9=Bޫa`Mn:TY&ۿ2N%nv̳cIqEeȫ!X7O!\Gz;>[d gט69x |I= n)(-1o&޿N~tHJRCt;hoFfaVBJ `Vo=`\rUk ag:2T„ h9<င)_XN.`X|B҂k۪pJ 1Rczc=N%A;H`k+F,?G ye<"%_Aƌo?e<۠6Χ^_K|^ 儆&E$a of"L4dBNE\'P'q!aFd RU#LԋMtIu*%i`;?FMJ(ZdAFC 8x4/E,v~xw?;I5ɨg_ׂgW;m_]3бNڋ'nz_rQcY$7u ) H qZ>`F'coeHэtHэ8:sS&P-sŇ(L@I2C"qF$"0ʠ?+2$3Lӈ4Ò$8BzSN)]'?%Yǫ䪲@{{;jN9dX)[d0.ᚸ߳pb]^Km=1NQ* @q69[Wp~PD,'YEG&#х`ty×.1&c_͕ yl,A'Jr:bD[8/`jQO_? ;ݰ(6m^|7g-SF;Dž;k*1`PPjRŌԓp'Q.aPER(`"`2Gq6#B D(%ho, ]u:2'χ-14C gG3s1WZe҂5tx,x>Y|+LǷfD탈`Ő1!VPdFREc#$YqG>J`#yk8e6wBww"(6#~\,.Uc%?rD%y8y׏yP=A3,1s?:;oxQ&o1:qeSa%{cw'LYvp~"2*,&l }VIDIi)A; IqN"*=j 4jTVk}w?iZ?At>F?5s,ma~ fr̘71NAe9+.~K~ץP>X⽕=PFO~M}uFG<ջ.)th_ӭ/5ԌN<+z 6 0%drqP:w¦ dݯ 7Y_/ZnX#q+ <] QyhQ𮨣O^̙og9LuҬ55-\M\ys @VNJzE>9M'?+ ˪9aΨCTa.诿~oATU l}Z_p*si"BJ$q Ruփa*pD2Dh*6\J2ơ)@Jk:,`IΪz4S:e[۞z&ؙ30)}Y4jO16aSs4;b<;=2 y,SÝ&dB̋隝.Ps`:{A>5NiDL.@S`\MƚqF/& a]Vhr@.D{_ wu-/򍪹L;k|7?<( OizON3\ 0/4KՎ.ں/DM? M꽽oZ;4baύMV΍͗3.g >jZ:VؕOAv\dGԚU|_/3v+4g3~qKWI[q|*"-P+P^6Ӑ$H0JU^;tx]mtM%ΚY}:vʻ}2=Vtu%u>(|;j3|@=#uU1s*V/0+BGACG7oxܕ+$;ܹ US77P1}dS=R;맖'#f޿UW3(xeÄ DqdFa2O0 EDfY*(2̄`0xVm_ aYՔ_5ĸapD1lOFN h//&\Jg1fb $W;5Q9K.d!X펛ExZ(%>ӟl4ն…D( $*-\\IZǷ%00n/Z R[kȃ)"OPT6 ӠuVE vE:\~)p{Įwmwq!LJZtŤ .kP T38GfkSgHOl4bYд;8QED@p2DiutﲄT@$|™@i&4hv{xö/N7/u_uZÒzݒsl:]ԴkrhG`46 T j%Fb#%r:Št5sJb%z_p;}?9Xm9^nQ/-JƫO0/jI@֫Z&D߫tqBeW;տ"(DWg/(|:f7ļ?MK!. W9ENnŻa?rO N&ԫBF&ȫjH㋷WYErէ$8!L3/ZYOWz2`6e^BE)!& % g*FKb* {<1!dV=Ծt)\R_]t30AFu2Wun)BFwi}`)oGZގiy;2mx6)hҌc24KKXm~1ApiLb I4479M׳UGw(1plP6-$llCݐm8O~8{ɍ`pf7\QĮKbu[]G7n0GZai92( I8Q"aIʼn*cR3`q(c,CEE8K([:׉v:*-6+T{C7Ե8l<ѮtSۼi]aԞNrK`rf}#fS؜l 00BQ0QBbB7eFZ}e(NQQ{}j #=H /=^oڛ,9'Ô_1`B\?u,E8y8"XpF | ԟ_/WGR\\?ݩ#2A3,1s?ߙ}xVM0"$o&bLO8Ƙ߰8g0 a'JJHugR ]kW_$P!KXJ mqv ~ti:F8Qvxex:^Z DP->gp#$!p03IG+HTJE5ΙvsnOjq6(oeZuh,a?`R(q6 <̀3_dI 8>`?.VO=4˪2b GK#SzJ@!R9޻.㲭l+n5x9S}W#ތHHtZTA-1Z4=f%Zɞm\KﴝBm&:ív47RҬzt;zOQt)3'稥ƠΉ1S4+n`a;yd ; GGwNet/3!'unHҢvѝBюvQ\u>z Q}| 3G莜n_{R @hzc쑃a_ tqK Hh~ 0In,<V9PĬ ΣN_q8M? 3gHuHn9nc n͐ Hm(DCi+%!tt23]m)j)a;ҳausu+SJB_i9˽!xn;Ž )>@yA 5CW(g:j{)w\p.4!M0-sCC\T ҺPg.~ЛH=>%SV'Zk8x?<Ϋس:*ub3d󏔦/zc(B#C0V6J*Krr_3V)MZ)d{`N\"ԋۢp Ri"%|+6UezDWƪ\Sy$S]-n;A%\%C aw~\8)XHOu˩w׏v/(r?,*^TYd.+WsQhe ,7eM (N׀^֫Ei"aq\ ş"%RI5roKJ2>4>L/~3gp=Esg Xj)1 44## %DLM("JŊ~x֝LHxM3YD$ ̿ ؼ0 xNsz +xCG8/7&PSi(%'NZxnSY\;CgZf۴AQ6Ms*۞k>S9:hUNr=kNKtJ̴w@ 8A(~( *ձPYC3NpN9F# 9{v7 }o =ڦK ppBb^-]]VD'9 ,5df$-%-]&$\_c!PCpDQ|"9>9,JBƸl4p9"k9r'1A0ERScl̛gɿ-hTL-Oj~5W롯칈S-Uњwp.ox|8  b>LTM2Z4Z܅o_6}xL$Q7?jn9G[7nhmںa֭_GS_D$ HDF tбFƜk1 #*"mi2m5ڻśL;U_]{2g"X}.%FQOub,| ..kK Ղș@&(heX#i gE@(#%TA$1Jq% cc}KlkCb[:(3L+5 uĉPD\GR $fHQH@bxiDpٔԒYY +OLRIH2-e+fC1 se30ZhktҸ+j^۟_?3v6S% mA7/e[$ozao?z72R==7!{E8ƱZ+j<6͠g>~3NI'Z,_{q~gfdb̙EMp f̎x//_;K94FVb" TgB9dTLp]FY"ajQr?AaUC[%.׹iV,s u֘7`Ptnv7di5Q@~ :)1>[V|'&׶l1&6-1^u' `' گTO/ g {^XX2qZ"4J/Uی7aIo[*#s32֪XؒI)5XX Ņ=j9 yZ{߉R<>,yi DצYE.1\-Tr__8)dҳ%7,LkoXk\ENW[C2>o2{'E1ҙ[ˣ^$2|쌿a1Nk73FֆM {i ^]~+#cOTM'Ĺ m̚$J2TʺU^+keK$xKӛ_^sh/ @ i}`I3]盃YK!TRPu2m4 siozex.֡^bɮ젖b< GJύTc}l 1.첱~A8ME"ٞu1q )A P2<_#A%#`ys nl9N @޺p1-cuu/J3*^A%9v[#e}ʩ6_Δ(q38dPH-9'FcMpF T"&a,AXWG\>">fb̛gɿ_?EVfjy?]}R(Z}eUѩz^)fʘ>|,*7 ψ 36S"f$dGXL!jFxj97Χ I)Ŕa}ZrӦ+9Rq:NJ^T*d`xzwSݍLSar S50.ys^qƿ\"-'Rq?В#4x[ѮpGM l\{IۺKs$4qlfum./6&,$j:6-bshshshs90B*|x0a(XE$0N? X||^uJUߓ(аOcGk`\rq;PمQj;:( ! 8H%HA)XVh|rY#盝ٹ/Gđ,02e6ٝ:B fHe~nVytsHav7owZ_N?O}hzt3ú渣iL}dǦnǫdLWzҷ8~UytDFyQq6iZW"tHY쩐g4`A_jssmÊSR"3(@m7/}غ$=2uUwn95)PMu}PlljݧQl$rsиb"ȳ|1{Kf[JʂQr)nn&aF9 F6qk|L2K?ƚqM@a#9}7Z,_ NIg+SD#!%5 ZP"VC(TKcx(`8JI&&t\GJ,ίh!A_ٛnBr21ȷ$WA$|@I@"fSeREf]AG<4& F06j#c1F@A0v.11TKZ[+չp,V-NI} N+*3Ya5xˉUuHJٮ»B1H@)oCsRғ ;~ GxpyGfr,GEz+h9*ݡf/~49NX3%rƘw͔ ݩytnzc\o9=|lcn;Sq7ʠӍjt(OUx@홯X?:I#gga-֗r[%Iz2GT?Li/l *exdJ0"Ͽ8+qfG0;#{Zt'Cs".!ֽˉG#P{_9tLS!\n-*b̈́sĕ>:h?"I;#P<E(,s=!hչj̷̡Fy7Q&0qT10!ޕ5q#"qC76v"c5H3d Bwj̗!Y?B|&TIj?k!`fͱtĔK"EСd\*Kt@,FeE4>c h6)îΔF;)0xC+M4pwec0Y'|*X.{Zn_O!,xO?-n2_V^o,SKK+t(gK% F,.DlycYNJTj1. Iv8yࠨq6`˙]H<{cDS"zicNyz gw^;iJ o'>2fdxykM x͢v.t%%/݌FWkad^ #{kYicpT sp,VLnP%:gD@.1() e(-g ꖳ. B) W AK :d-QL[i@ndJ(lі!sڀq"S-QX!:*0rKȻYkL蓄n>TSWv+|**Xj7݁{uxQar1*wه/Sc'l]T?)<߀&{8UM4t`nTZ\-+՞oS>V5(Ū;/\Dsdb~d[IDA EtrhFZ_-Tօ|"%S\ǹf[.RD'w&m(BR;[eݺ/\D[˔Bk9\/7 Sv=c<\~OyG{ۗ\Զ }#;7dַ!7} ؀ǛgzFxåo4AKSKh7Gq4Jh(G[áWF:~r5JBsKB~#&ɦ^mO6=fŨǭc]u҃ftu~Y/v^uN{W$NYj =UX=⤇?>#/7CK6P-?#)U{qZ{hpF%]V/@yh{C||zpwOBl)?[o/\s\`*-|$} ؗ㧫t7{8h3~Hib{Y+EeA0 (|ߡ!_6 :@C>;, E+sX +DMIz[޷Z)ڽփ&Hu$=/ٚ 7Z-?5;4oՈ~Oiébv_oWɑu!aoZdIԵQ,uW(2x

. :@һ%}终w">5Z.I 1T4j u - *A 1[ϟk)f>F%bTPYTo6mnv˪m| F}VS՗:1/)>BM#ŭOjV7ZfP1ila ZRXК`etp Ʊ9DM}Y,h潞Zu#Hd"$H0~^>дxYeiVi֝Ɛ}2bGLLOͪ凞;j:7U<}AH7ւhZX7mO`Foa 6\XXw HcM94aڢ0^C}84Q2̒aPK]BA8PpX*4JOeW/e>#Èd"b)郫+,G?I˧ ,$Ayl%$[14H#6H%k<GYu2PeB*9HA"W6Y?eˁsG=,l|s01%&>P@WBrh '-V{Ϡ+@%A9` 4Wr֡R6h@UN–8dm$x3$y(x.@3뼋ԥ@VPk4/ O!/TC3:EOLJ JЗkJ[)˂mICQU\LJ)=Y,Nk_{PC&.M _Zk]\3odtZ>~WjEI$Ӭ$׈ylmUaA*ڮq$_IzrcPʲ f+Z| \NmKI#T^||PWd`ëwsFOŊgj0Ow>Xqb#IJbwg4D_RC5( %#,HlZ+~avaJuUXEsHK|ਹi4ĵ-f;՚8 pD+aBXB*-|pDjvJcJ.| *iՀћf_;D-0lC0RFS C%-QjiAr@5g C}YF}jJxN O?#}bIjКY1l؝4F$v#OaHʩE Pd"B2b&KK(&?# e)ܻ.ԳV1' 8R_^ÑB^jd4Bb)aC_t_v7<9J4Oz5ʠ=P8!4gP)+*ZuHǤB orhBQ(»Gll+m08j==1Rg$V X (v>iI[ )0ZW@lwp"~C44"F. %>fmj?40lNʪ9:[Rk?ZumETJ{D*yjkh=Jd>.wW۩f`i`B+QX7aB4&T/tMe_^ل -_[L6s+i-;cpqItC STʏpMT/?Y]Y6L͔Ƚ>uF}*Rj/z k|&oKo.caK P FSu׾5_+Q2t-U:#NR<릛n$׵+ӌ5kf!$lI&?&shQ4z+}8J-s 1Z [Dci()G60&zXP^F@bƓT;d(A*]r˜H1rWN[ iJ$ٴ#) gh7_JZL"1aճwbv=Ukk\չf=[Ͼ$czV!p:x)Mkd3s/bb MjnŶLg-1wVVR֪`'Z:ZzoML Ω0[- } ۚXQ@t[.M?0+o q+2!?D ZyRV)4S0㤎E)߱f-upP]SS GxQ'qY-? n,'f-L\(-Li1&NƠ*+ 1~RA2S e2xrPG%ͫ@)=h QJ !C( VPfmE4MTGoq6g hU֊>Lply{i5Q#.l05pYd #lQڷ:G],-p#i‰} ;K+!ʜV=pv<ǒ|4!/>{hт`.p'!רMB( Zx*L?3ɷq9o؆#Le!T•%$T7*D͉4*dqIV[I+  S9{ZC7;{#hi=R Hwa2Y9pO| g9m@QM5kmy !.˒A`Y*畲ya$e5ceP^3G" PVLhNm&wD/[#B!)P*΍Q .@Y:%@ / %0A^RDVRگS P>(tiK񭞺EQE\Po*e95d9BE([;Ǡ,}ւ6H#o,Qx☰Nk&zfR($F-H+ecf4^=\;U]HEJ[s3['%,?8ê$~cJfR_OM"s-$$$W/pw)XO žW !h )yMtCfVdX.~7{oشJ`H*wD@Q~80o"1f,) )`)J ߕ& -*kۻ")n 檣U|u^QG[Gi}~g4KZk;5~ˏn0GD4yb/"xI-_MAY/ ?+לmf~fk;k0Ool𰚦~]=y%όb-R!`%T#&#_UMm8F}p"(y|bE]]¿'V~yժ/]~/_V3 ZiK4&p%dsZN %4m5N]6]M&X:Fw(UڶBP;5 :p-ɼGLo6BlDcD; C߫!'/¡ -EXkӪP-jg#Htl;-؝j(W-ȕ\t}9֙j;5[ K~v٪s{yw/ߥ_ǚ*,+7>6<}wӵIz21ƻM{%*ͻǞez!,+7^6#WMB 2w+m۔ 0w_h݆D6ŪK:p[緺akTf+6ZQ`~wݵ2%w`5 16(7 s ۊxCv 覇%AҳLlQKmBTjih{)v%6RZ%VmG%V{;:d8QV098ҝ d\q kKPbym؎XI7ְCl$XB6]3Kȇ)# %.PJ q&DcVҬsX.(p1R:QYos~Ygt Srqf $ΙD*ho sl2Ap24Ad"4ʄ7;4x()G6nEafI:$8i)H(.AIjBGg t,e2 X*]mf4M+ΦL!6*1a [8ptm7o/mvFZFkO\㪺*%ȵҨ$XT@ܜS^9gȥ8XbNo T*VbX<2 qoR"J*-Hꋤzo:jh#'E~ukAp Sw<]LE,<WVPĽ)Y`LoYrް#'ssSK՚a)JR AK/5\4).㲲D$MCIU{ApZ Y'؝!8CCpBheͳ6rWPw&K֮l|7;u1UqKxW.c@DB^T`]?2s5{*[֎|M'<2M8D7XfCzjy^[ᚽӐ[kuz2Vp݆/Dwl$ڔY0b`ش@cSް4V%Q!m$ۗ*qZvP ֞)}dXQT% hTmR:q JjynhRC)C`;0&kssLLzAlJFj.,|sUecn&md3zF;Í'SDwM@{X8բ2A,|ŁivjcS>ge Bc#Dž;먗A欛~1!}˥taޮJɔVW '87q-YVKж)D9BA}+PT[d HR"ҡ:@raJk_В*+aV7NӔ ŹkG*Qj c`wZm&kaMKk~ypB(~z@7 I6 Q(Kltd J _Î3x'DG[] sdo -*lY- VdCG̉CG-PW+ ŊU~! P0ƽ?NxÁqvt;tlLߏ0_Z,ˀAT;E &I5Xޥ$R:Jڞa)Qdlw`HlYB;yXO7JiDkz*4nW͗EP{h{VPbdžr0@G ޙlJ(,ژjmpu30ѾD DcX0;m.CkX(Uy\|Aݗ`pi~n/ݩY5z!PBEF;/~u~Ar 甯 O8rm;!Ms;~9,,z&2;I5@w06ژRy{LG֎5} &cxVPAVVF7S0Nq9hJ]$ /tuK ȄlaFXNBW(_UP lؑgU|dA dF{rBSբ9TO+;d=i+~RO(,#'uNDBaFh15A9$+njgj2S(P[al!Q4qy"NRp4oMgjG:K۰fzX gXQ@c[tTDWFF""Sb֋<]0>߯0͚O(%kvPrƬwU:rjdއu|u P~ DHg\7|vj,U:щϊ`lXz Hn b_o9?_vu}E7\ggJtf5#뀳3~xmf YC/>>ZHƱfoYt]\n뢝ՁmCպA_-v]t}w\}fFoVٻG+Gv}|⿮ͭG.Պ%􄭔 y/iCᅹVm(s"ĆZCLdϵW%⇸j A'y@ٱ ~X)LgA B.MgQ?)#:D3uK7J0pqx7H\9݀/Sl2G,Ղ*+]]A_jIWjR g# ߜXīMhz%I$iܠr$ᰲ4l]ӷ'HRU)LGsZsT_RMݦγ&ބ.Yi 4кkA{YaNF(ۄESi w][]9sT_RM,ymBhf)$YsR?\yyGv>($gK,~'rx2Ly3Y(ؙniq zlX#N|+%K1awz\^k}Wm:# "rlF!-ꤝƾ= UufG6e*jUs7r:}%jQSٽj!:;IERk4#*L˛PbtsC)lRef׃ qJWDN6'~H7sm"vu2Ŭv﹩!+jMc6 +jF;_uCNR8\h' <$zEʶ^d"^E4(6')'e H'*AF8޻BAh6Q]1q%J$3j^Yԍ,ACitmڵCZ\ׯY߮޵w Ϯ\_Iğ͛Tn7 4n,h#߾xOd^T \\lE'`Z@v":m77Ǜ}k5)bcm Z߹%b|@%j [>ȒKx@Jҫmx̩+Qhwj021\nӔJ mA{s6gsT_R.u9J쪥vj~F@Ml~qZY6?]~Z!v{4V1q6y()iCdRۙ2ĞŪ/Z[lHԢVKa & s'-rm^+Egұ'TNcuVtQΛ :@Uok >n|^ ]t$1*A;eq'^6o["y?<MwTNd!_ٔ6GS?ncn2H1ox/ r,`LV^/݆/Dl GO)u ޭT)S6)λf$z!, 7J,'XB;uh˴/SްR'jЫ9;cGm&S}_KHR(Ke(TB#OJ1Jqlp T&*'\5u>6}5%\J[$$/٠;z>^-@M7B5h-byĻb hgHPuz ,ļjا]~a&ͳg;%p<0aA UT?~ػdl9TXƞz5AJQM m~q\A5- FiryiT C\r):k8 }*2%|dH+n3FK'Cm!WOÔ< DŽ7{$,{|bVNgi[n1q3EZtv28ɵAq4jlBNQOЋ5ƓW90!8? U&CSu y2'>ea?,DtgmcHBOTLvojeb5^),,O5E 4Yʉyy0gI6+,}dZVklj+q1E$eGQk@xyYX+C͋Fe0J)k (5Y͢ GEj ;G$Љ}N8|b :A)K5(1$8ڊ,M>u1FN9ATc jʲu msm=Yjwb',-mSJY`W5g[TRRw"Rq3;eTq^B aPsy|s)"LJRJ )}2Լ (DI)m"dl'!P9RRh.%r'0)͠tߥ0)<;/ez'!P XJ9 ƵdY؆+r9/h.lgeh^R_-O1/k:ojZ:_*- P,gFVV❨9*T>VZ#*rWQ%)WB2|b%Aa "E5M\h3+Q R* T 2â1@Wh'CPPxXDII`$W,c1(1H1a2i gEVA8F$ jJDP^͵UjTMވXߏ"[N,@iHpD(ǥ)X0 V` +2X b|9>'g$J용s_HdG_?9oܗ)r}v0 "U1hܗ-@ P}<zѥZ9u$4#Bm?+*m1a"ϝdo1a"`"#WP%?}/Ȱ~.ԨD"f5b$6Mws3OY ɰg;䣙(ŋҰ]+UXQ%rSꖧoj vn p{傤u_5VK2NL6`pr>sM"r@FJI%BԀQ}.\`# ϋ+OL TX;noo5y pFO ! Zg +x38ô^g;Munp!čö1k8j|I!: "ָ'kCWሒt*!ٖdHSUl- SN#$a1O#4ታpH;#YR(!Sk1i)WT#&*CHX61"H{N2F btQ|hFRKcMRD%%L#;wK7OR1"[o\:r@ h}|Q֏HOYgO+Ρ 4VEm0< %DJX.D k%6!ScKBJJ)ڈRha[劷 Bb&lSQp$@$U\BL+p/#/ek~č-ZW̕V^G99|e {LJь]/';6`k.:ZKC-:lFH8$AR%pjR*m1k 52*I: gı3#X'_$.]s2Ew<]8"ds3q_?rm:+0^ܿS,?Mfq&M0J:><Ÿ{\9B<ՓVewuO/6}Yzu w.`2OGt<7zwOϳ/;࿳x?ft3W/_z8sSgNfx8^E9߼^v'w߽fe^uZ77"uYX/^?9XnSoF~=ӽt=){ҙS_g(<&y]wRlOzg۫8j:j4h8!0Wq?'=xzwPMFWs@X'\o|2O2Lߎ_‡} ~ٟ_?S'?vyљ㇡I+2I2:%=` ˆW}ftϰq]G]yП49F] (I^vSgm>^'wE~Qe7/<:>~;b;?>yj&v~(yi?KOQ X).Ʌe롇#GUWXB1BKf0t~̖l~9㉻k'57>hp5̞ b噽O_Eo;¸;%goo' S5~ݷnݻ!^_͔)L9|w~O5Y%|u7سBR^iqM @lG͎㲟N!nL2B! D` Μͭ6oRɠNG=v\ oy%Ё۠8[[i(4iLkG%?yZ5;JB $QZJ(\"mn<ئN$4BZA+1U@/ߐo?x53[U_]]f2}z6)i `ٳGd(x:rwޒ.0!CdwOavGɔ#[_Pݏ[Jۜ J k wt%Bg4/ͧiv5F`Jk0S0k|A0^8J E%LΨ`Rƒr!<#͹dC[nUf#].Fӭ P7CţZ}pC/vmf~v+r 4͍Xë?N&Cؽ1m᎒ZVGV~kbٹ \$joBDVPQ|G޼ oo:=` e g UooM xӅ׆hܞ!)GS5{4ģ~u(HMVAa4;CbðK9WUF`6c:F6WǬ@S\^UQyOW77.oZ 5!6aSg) @GֶBsR 5np1 6BJ⒬[2Dbf5 aaS>!< yHjnMU^ȋp=Y Rg;j&SeTE>K[gɴءpdR) i$Ea1ɔ&P6xۖUCo9Ey3Yߐ+PnDH U:Y+Y1GrE}J*Lڝ9=hĖcWa%wIu*W);>?ܺTVRHʾTRR f]Mj8ht7@_3:=HYG>%chAl>C yu7 %lU!004/0  EJ! N8Ж p<6D s窗\mڹQB*渡\! !  őh堔̯2Lɴ¥sXxXD]ymJG1[lV@1Y-$6I!CI=`ELZ!)PRYbK' m~g&BB%Cʨm0e}T(X P)O2"d26:ҰwK%v"NVGJ_(^C2AHLL*ؐma>P[50}ofO Xkl ogx0̟呲^Vs_8)t\jsC4&Cu#CDYNx(0]@SRWV:xgwUgIl^ j/QGl镊VqZ֖pÂ07! 0Õ0^>8fjmo i ʑhhLX@7ʱthxGF\9Fr“ȵ|>:IcBtP(Nvo28 π΀8FR)TeLjwcdXy_H+"-UbG3*H:0'0دh^ti6NPF\ BqBTQQ( Ucԡn^ɥP׷BI?r( ?EAqV0ؾN3J 9I(νdcBo_[?&y}}gBHKşZ߬Cr#z+ad*X.-QK]4ee SJo$+?%$UJ[0It!K 7Hcuq`MRklWE 7jhD9~{Q7 F8),NaCp$!*lUqy,[if# 7~z :xeO ޠY5@|y}Ǐq˕VR y*U[?o(w`WC|3>@/ Jnr>h1Tb kEiz~{4r >ͬ9=1zX܇IhP߭1gnfY93"z'!fE5v$C"H\idepcE[tUaD+n+-eJcxLhgqJmv:sJe4Ui`9z"Bplاn{eT|@ eqVk֝0iw&FC6O2L,3Xӆ^ O$JE#XR sqz\I|-l`jhGnl8V[>$${v9#x) c))ez,ޫY@⒪0 ; Ü$ߏjam$`kUhp"_%M-:e0~G,,tcwz)7|#~Œ b'"rYc 6VZÝ<+E$Ŝޑ_QiP uHۈCoH{a3&9$Y.蔚Ru kl ޡ%qⶩAS#uHJ؈DžIqaROFODž eX&z?.uH;.Ld趓n߇cgmiXZAzh%ɏ$8^` -h`\A硛4VѩFaaIHe$/X2V-T菎Tgtړ)>1jس _GbR"i؝jI(_{ϨVjԞiߕB F-o/e0+nM}nݲ˻ܽd'_]6Bι_7@)kMxs+-WA@^ψԦS^vW; <]Sw(4 W,RߣY7*[@Gn3JY5(δh60ehvv 8.}g8D&K$0C Bdu=4X3I/O(_rZZIo=4Wr+D ~Ge_3d>PҧG$"y#xHQD} {DoC$i|,) Q7)jcJ ԥxX0 { 2,<7*s xb-KMKSJ8u"Ry!DR–KB+ T(U]#EF&9,5,jsN ;GbwV do ?WLqtڔOqt~Sj*3̾Req- }"_}YBp\,qX-?^O<xG-{wQ/w}.?~WUŻS'] `=#gW/Ue=wGPyȵ| uG5r|?t%g(ߞ։toCźQx/w|6اgGc4<7'#<-}]I~ғRBⴔp&1LBK_rjȳR4 IhKT_R-=m-e2NK 0 -e2NKk9ꬥ'i)s!}U߷AKSж>ܾ~]bztPX2jխ~TZy U0q}+}ЫW׹11Ny!U#J .%uu4טV R^D#7~es%c|㉫C(6=@X`Eb2سU$׺W|6aB$"y "`LS :vOXF$ŨnJ3J5'q*djn12R M 5):mpRY6L"Nݫs- N Z8Wr9bH) .k B>,\e,f ee˵}}ys(f7ߗ )ݮO"o;dH2\N$r\>4p8oo!뻸-aКܸ5____߆> ߻kA^y m^hv[>js1[1I,<~Ȓ #|$sYX}ve:IE$5hSAVsT 4P /HB }bJ+k-p(5P(~^>WUl57[g1uv6ARj,6%YFdWeI6,!nMQ0L"{)>X"9^yEd)2! F2[@aj*2\~( k#m>,"LXLc RAQi!ot$%%pW[Yʿ3u{h>4_|῾ۼ-s"B9>>vI. O`}] N/G[gTVvwkGǛfMD~.T|lW'x.?Ya\ܭխ[sypcPdڼA(޹ c>hEOo[_yhdB ?QR DS*­>vӃ߻UK bOhNMu~nݤ)X\ĨN)w?n_1o+|xagРH*}N&x#e[v$J3 hi.䯘Yoޭ=CwKa!pmeSM;jѿn7 tھw0K:y0o-MM ħJN<2y&DTU{qek/>xo\RSo?5lKAXpb c 1VrdrS.B\2+wyfaSI0y=Y)"YX)8+-hm,9e"YX)D(Y9ZY *Szj+⣕^bgX+JҒjt\/J)R"i043(bY0@΃\)-{D ĂڔOmTVcvkP:\<Dh/Li˔?B;Pxǩ^m\Qejk{Pm_>~-5|ud:#kYB5Q WS1zѸJZ(q%SAn<itCCףK.Rv1xhEvm:8W d(F$'gYƑL)R"2x!QQcկ<{oay yMJ7T6٭q-f<x=n =O0ۇ䡁Bp~ HlN i RB80Q\N[o'bMab|"fpnֶ~ycDabYAe]df5ɀ*Q(ɐF( Ju*jNUa=+eK峳ؖcj#ǶJ$RdJÊ6XQVAY89"/0ɉ7Itd p,'Jn @ۙ;mEk[4R4[rKtAW 89&.[\$55r=39'&5jK(n_|uzh'J5Tնcj[L3!5:eEc[]hߏF9g;55@_ԩ$LЫ׈p R80!T0{/#[Mj VVn~z<-iRYvo]cPt[ fǏD & JdM1WҺM 2"mܳQ ,cme'- Tq60́aűd-Д.rj-]w#ȧDbF\ s:Y/eyDm,*rsF< a256}<+pFyoUN*)_ͅ'Fua'G ͓Wdʓ5* 'DN bE oaMu\OSn֊jS.9 ^'#0iH$f08PI_Q]w~5ԧfX%N>-4.t#%43捗3}֧G1O80Y`g1E %!1Nы$G`ʨPayr79 >,9q RP[o!08dT;SC)``)u6k?0DQZFr١FhˁG,#T3#ne{Lし8x!׻TsLu벭EflL=׻T3Ahl8@JqX>恏U! g>R ^(E%)O\Z uwХ%fpY.D?ض3'*]n Jd<$ijed^[ov1Uc^ .ЭF>+r?)lft{K׏Bt8-xQZ:WƤwX: <5BW8 EdcLypt F1!:FKҙ!`=ejm.B~%I%QEPrE#>ȂH2T}4cl- a@9mprNEa86R\db z]U1.QzHBB8fey8e9vBp+3B(YJzqUm?.7aaqNI6 ǰ g&Mnrk3 -NUkˍoooo]@%85iyᗫkǬ``1(=k =`?ACR`"f9WR*ώ:gzc:$|9UZRv ҚVvQ<ڜd P,*>P(]?Op8 +=F.>^lWzjhm$JH TE 22WL0!6>TI[(g Fb"G /@8䠴U)턏>]OHh~qf>Y.]Y܅n+0Hѩ X?٘rňm=ӽԫ?cW<G< H5R-SpoX/5`mc~,rICN4GLEw2 g{ϳ,_v~J1.f~`O@*D zi%%gR@&ǟك`~9˒ջkѨkOιV.6t7_.R/v(^w;N;1 w}v]NC[;]n;]D@Dc@+Zoݜl+K.+D*awoн[enyksF1cT << V+1HưŜ[J B+Gԡ 9BgC%3Z2 ~x&[҂9Id{7ޓo:#oP5\ "5zݲ3g~KR'T'^֟OF=JUNZ748W{WLPx+nQQ~\2 мTm#1o#Gނ%ԊlKPwa]x"fzg<"L!셋7Wix-[ R@d:$W$p}ǥJxUFwNeZxoA3E8>cs+gbH 򊥴ǖx$)ێMNubhe )>ʒE_OU e r*'_ygĶoR#Dv4EP;n%2:ɮG9&,FZ/jˍ%kcEy%h xJ(Cnm4 {%ƖcH"`Mrs#1QCV(fHBN2E.(%[.dJ~ܷ$UL TYa%x‚AB*/s eJ!'9 qZ^+}Mܛ7˹U%4~n;O֣'i$Ra3μ T?îl]ђ+9InwW^f|gJz?s1&Xh{5/k#S)4? %&kt<_<^?O`A^?D?t#޽U>V#w ͛J!#en1gL}3rVz]9j Ru1bNT|sF?-28s v-n;rI-vJ*d']i'Տ{6^HzSbobzjD Ķ6omQ%酮lqZ.vY?<,KYaTĶ 5iy&)2㎺ f  kExǦjtJ Qv+D i[jQ3JuX{M#ݰ]xHަW^MI|f "Nr>wȘ`ama4; ]rS@^^vx*G_>d}qkr&b,w+No&M$7a܄IrS$'rv3~X {\@q-[bxnkF_UEXHp ~مn=In=#52ؖF_UzLI=?NK, vԐ4蝶_~ZďP,R;ZUW߿I0 O2ӱypUn>N?4OgI3S6E6]xSOU|ZEqAUY_5o^|rd꓄INV^/x t+r~pgp6#ieA֛Pn*T@Pٚ2HffD.)@+ـ\$Qϭ1/[z)CR`D6AL|Z)adz JبkR55"gu zhAzM [붂>1Ce, +LC=Ƴ+ha%*ҨRG;!׋id=Whnd,U0jϫ[hG\44(mU&7LLhwL4",T\§Of ,Bnaa˲Uk y.:j*0 PW,ߙ~#vpui4'*L姁ש۩qX+".y~L+;!7Q^Y %#xpqnh~{/ik#Nx!VgVLy*p,Gge]9;)Ѱ}II] 5e4'JbUͼ"E(6>]*.ecF ̱\ GIDfڀP3B^{5T"B^ir*XKVjYiqVJGH,㕫+xx3/E@v1NA(#<9YQ3*cBц%t @( 7<nW$=Z rkpxi811RH>Rf`tntLRlF %S ͍ן3M(:zh )^KCEx,7g ̏zX`=Y4BI]xtD 2jpZi駴 YOAy\5x ٘f~E"iCcP|>Q;L؅jDv%5y8 s-u J(lBHWe@1':l‡N|r`><-QrVkZi,GAxNz>H吙9>cNeP$&ڊ$f1h+ ވ*zOՕ58*3 A0n =跰tc46۹K5.⚱R8H7*4eOCڊ˩E1rn[lrF/x^A+`\P͙Ŧx: Š-ea-ʹ!F 9#HM!E"o[Ewb؞#_/gB G4ҟ8+0/ApQ g,(ƼheyڢqAYYԁd:ZiqwB>0>0Fj#d2+))ykYj 3QMkvbE^WښyHZyU.LOQk}p[f{Z{t`BdYSm1D+(OlML)@i LwP,aic8"*@ҸyQtLYHmX| =jP48ziyрPUtDRtW* ʁH(Fˤo<`3:dv뙠_sa (Q^x`9QcL*ܵˤe,{9kV 阉'%L5gZ)gӫ''À˅~׬:y!p/4NJL7L@y啇}`GYw cڇ|Ă# X;Vc>k?7nŏ+" 76޸\N 8ja#"iᛣ%7KQ{3W w! 2k~"#y)Er&8ÃNv_ۗB ,/21VDxk`gE(Y<қASJ1L= mSa+]þ7SM&HuTe(}Q@oc($wz5Ʌ(}+;S-(~gU}nqN?.ZLeu5MNBc vfzWaxw~?/Wd&ofWoB^uMn=9eQ˽ 5^F\mKӂ^+%2¯Gs$$;N}˰ScX,b 2jlh7([_ rD;JLO)*>v_\݆/\D{ɔƣ77(ld=Qd̅J<]ǶWۭw\m ELFC͹) T/_'b=zlǔյQBkhمw|#S&)pywW-b"\NP)p)ߑ)pqq_)pșQKpj\-B" +O`)Gggj?igNY %>V061)К==K!c\7{P٣,Dl+4ك;ZȤfj8Si൘*8SC !^;h#׎DK2JCjZцd$URG$M`6)@!ˎo0gڸԇT\§Ok Y=vɵK~b& hD*$'p,}uUXq=ҪI4+i3ӤjOrtTi}|# P0Zd,w! b3e*O+\^ie}xWV%g-2h&zQ%ٔ,݀fL;;Stew;ep vQxg6i(A(7!:$~nKb9,Fo(6Ep! Ƙ&Fmd1z-NGeGX1|m篿c^tb0y1PfŌ>ޜ^K~'R-y߂<}A緁;=LT?M)`Ht~Ms^L'SF_r0Sej%WejM>nR)+C*0GbT_Kd 0>$K sX/~@o5&[!ќ$CKЭv;6W$-\gJ#2p5K9ns+ҹݿ[t;ӯzG[[ G̯GXFlJ*THZ=S博Hɭ%2Ϣxm~ V=M}dAe,~21LpIlǹʍ_ڴV;`!ha;#sۯ_^0^XHv`_J1 B 2W3n 03Nגԇ^9<ݒn5j~D:+ Ȍ05JdiF; jeFW d!cAU*]k)klBB8[Yk Oc-XHϤz{@HNZ\M>='4?ΞON޷U-MW*Dɻ|{7!Mgoۯo_C󭿝]#9Li[ѯ$ͯ+n|wPzC`|0swfl3W+'F`Rhj~zrh!9ZIa^M~q÷qC驊$fHs+a9sVaZOܾwwr`^35!ٶL2j[jH s- hE ܗVR1y]PhGKX>w h{rf%,˫Ϭ\ xYe<;kwv9 ;p(-Zlu*ayT5^Nf;OH^{2,AV;:vifP7:@LxvCꦕ)-sk2j{ggv)?OWT{c nZk& ;ki{+ ?2ˌ?T6ec;&bo/aGJyǹaԁS3[h_'5E{B6?FzԷhk]5rbFvutMݫ*\/B$U2414141v[.KZ-Ե2CvSG+ˀ!Oq._Km~WI&*%ê1yܙ5L پk!t]Ȣ`ԨT-#>H^7Q:(cJ|]sm hwn}sp{ַ7O'FR=B )//ē'J<)Xf 8C%rcLZUo{Հ_0cx0r+VʣI(P1r1Ȳn/s>L;4SiqqCG:EJ rήtԨ^DpZnj;%{laWtop/m|kwsg#}H4ہGFvofy09kBA[3Hό n{ ןE=35}/.ӞRQ& gm}C]e7nm]}t$c Jwb˓:C^/ g.u;1g;mF)Ұuh;lȏi<-|Ps;W+s͏u0N8JR(?=~hO ޽ݱidي1fE>ۓ6UR8]ƧM>lY1XLCYWIJ~8%?@80Ly~=DzE͚c"zm 1x[ T.>.99AhuJi0)JhHۧclV0y%k?9KbB1E.i.5LW?[f)]/av 32nr[K}ynh[  ήb_͹ŚݫMbXKVaJcV M0@Zx@O͋!zqn)B;av<5~$ZG\+U[A"㾔^J=]_UZagA75h7SUQZZ0-U_t؆utlx!.VP^вp,J_0NzJ7.+1zq @6wR$ k[?"$ʼD-ǟZh"ԇijlk75/ètS;ᒉ"頙UNlc# dNW{04ָXzvoEʪVtt5jQ1ac| vŒ_􂾔2cQjDGhфE Q Y<,1%'eMn|h}cRC{y  Χ2Uq/ĴDSڳG_YDYH#ю ~o^Eƶ.#`h;>?̦zTݻc >(B]8óޕ(TK wӨ.U#镁Mi"7Dcki֌TRA$,2 ص19a|~+m`C q HCgV}jTڨrrTY UV\;*1feP!r۟*xlSbˆt7 sETj ;y Rmf6/,=I<(:``a6_Jb`GJ l%HzM:xr l?n;:tRXг`hcYDF"S6@I}!ۮy4;?)1W dlNXuGaX(-ڡ؂7UY-.ӒuULm>C"JLRd*˖}y NM/bɼđ[ꜧo5XvX t*_Y=ÊXs}ĜZ\\bc.vU8H(VT:^[5pHċdM/hj> QUucH#;5+7xK&ۚD>?0,NWl$^-D]]i ZcxnR6Dn~_3 xH( $5,x*3 "ݰ]uP$uy>y)sa;Ӭ)"xI2ljY/m1!h4߆{U+mM\jKp!{{֏4;ҟM#뀭=7jiԝ:.8ܬ"-ۖz*Ɉ ǧ1ǮENK{S۹ZoZUM8_nh:p}Vyl!oB~J0JU:k:Oi%.W?h~ɠYҢXׁmi{-w'$wJ}?{,{I(PX C_VC,s܊jQVAl)WG/⑑:]yopkʿq*5h hĐB=jU Y19 MK 4i:PzNΨW7@؀7x{+{gewl耶vVRF$تiވz Π~J.xŨЊ1W-޻#9m3t&"z+VztAGeOe:ri.jZc;,ƳUSOD.t-ƴrkWp9#Q03pQ!y`Aʜe&TpCo G6>S6m-k8fih᪊%U|T< b:d/ NW YFB$k(ҰJvHr_\x^LWl!ȝl bR@Wfm%T/]tPVFpO2:jn'cȷZY}fEMDyof5_N:[xo`3z R序eݿ. Bj_?aݱXt:eWu<j}v%G-H{Ӱ\=(}m7 ٲ?T#"yl~Onf3)BZ^c3ͦ8/;} N 4ЃGlm,A]61zm]ocymvXX })|֏0Z""|$5vS^N"|ZKN9QpwgZ.|y)wldx*KӻqHC~Uf-8!KV OI"cH<#tcYwh,ỄfȪo0rh5k2! !Z 10d|KfV^=481vᒽapofOϞ*ľ8^GeOc~?I)j'KGLŒw/5ԘJpj[}iɘ,0ΨX$|Wb7G׳^bQAe@I"p:a]\y#i޸Je;FE;cV%d#Pakvj&eM #,1j,(=II,!W+ "~9ɡ2}/'b꽛v۞{7zA5JZЃW\o/\]?׆r3Ɋ@6ZZ/Z?۾] =/ykZqn64fPXpCg_R5]ߥAZpm"SED a1/eKДPv! 4BhXl:( lb(5<s_(5f9ʺ$%f^M-+ Ygmq^wuI%  n60I^GMhs0+2/zhBC֫ͥvIQ2IiO"PPYHY=AplSKlOflXngr^\+YOb҂{ofMQ6gdWoե{w1ػ&{$ f|e!RT^K|V H3վ/5[+'%z/1.WE8/rTDi<3$\$JIֵܓ8Eɗ{o :nE !JIrE9][o[Ir+žVWbfl6 *ѦDdًTRD:cKu_Uu]Bq xCІ1ȥo}3C= 'QÛw}2 o ٮ:SSqoA$fa #+^M% qk8yA6R|eaN&8t>`cܐM lIAs5Ye(_.6ȉ0:b.AK%|qKdhjt6qb;Utd8Lډ8חe:+ x)ЯJ Fh-=PJ BԦ1oQ1qbУ-OsНI;덟N*:h;ǺPk L:r JG/yV:ʄX ~g|" '"RûAZ,ռûzrnt]6뾾<~<i`Gi$"z:E׻Z]DӰj"~b]~Xv3"붜\M*˪?&M_VQğiu'1C; ʡHn110CSUs9@MMYE|fG#+$GI(e)8 u zSÜJ[Al3G_ "ᰃb+\!!$p(sI:Nߗu["k0idYdt"ʸi: 91kF3R$`kF"YT{|1ZxHn :hXxER&@k#Y0$^H ʁ upȭG]0EZ[%^R^x۽$g>R3ZJR0;Fjw&I^#3˯,Mz&xuwT#Qہ o̊ݏTkR|nuM]7Os]Sm>zW=utq˨!+#_;WF\ܾ*#Q톇|rx.&QS@e! _ B#6_ڧ UauZ_կ:"qngxT(%ۅ7q̓ûU_hfm*S^D>ήiU8W7)M{0]ByglJ2ׇY^4/bfgMxF0iS/5aCL sH iJzVTOY@^2j Vj7-h4":}?}*O[7nvnCH\D4՘v-h4":}?Li?ڱ[o]їTȷqӳ{s   ND cW8;A}iI,Vg7 ۱Ч}:gu)9%U-0pE^d|#.0W:QCk7cgciNly-b๭ؼEaڛ[R8W6:;VZ5Nj7R0h.Q{tA䢅WS&g,yv[ :$շ[ns?M飫|һ\zyuZ~4Ҟr#ޣu8ZVW"MdN1.s\Av4~^TOq7ج0K_?!27o i?#,$ŚzZ+,ǁ3]fe(mVKD2:Sd'x28%%#n ,TM|[nod:)w@*~4Hy19p5IXҟѿR4sfQBPl4L9m9kD"/c} )U]]@G` 4^!b("Q)9n("q@r'vk vb s3\bڈ̼ԁ eQHJNwîUƪY|<\mLDxı-q=(A0ymq1h[o-7mWȽ3Ek IEѳ30]LbҒ&u`d <?]Nnև@߅ ;:Пi 9-Xb>;Njq*sSX.|ԅ }fm:eTq kV x Cᯉ!hZ9=6%W%.?MV dZH5^  WJL~\:unz:&n/t(G|'Umer\ ) ħO}K9gDw$Pb25'o'jB*堇o; jf6W͍q1F2v4NdC"(B=v ,.d/,37SLN6:{2Ep^y߼6ݤ06bl?/0&[L"냚& RQDc:^hu*8נ6ud*3!T,y~օtMN1_& LL)\7? zE޲ 2<0=(LE@1zԷ/`-PY:nsk;:cL K$G"JYTv(I7]!ߌعy&i/7WFId҄:L% YP"f2zl|%lq^ls HʴօHj߈H琒l#pXW-v{k#تH_\-ƠErK%Ż)K1C|1@kNN߅;+7j$GRzE(4%@!jGŞ}qd,em zmhƴώsj6Kf9hhޗH'ăȳŧ1F.smγ VSEMn&dH}4:cIg_Ɖ[$>+98^dWIc{^C]N0G /龐">8Q ^FQdl xJIxcV/ߟ/]_"~f?_M|1*:I.pvxOem ȸ4<SZc"*2zlˑx*O-td~B#M 1 8+׏^keI$TY6"\h>}To~Sf>>flE.y7䒟:8v/ΰ04F mzMa)Elv#`q*  I{Ji ol\'ȭSpxf2W *yXoVrΏ K'hɺJ9cm5O!H57+d%NADm{vuZ-9ݜ>9uu@=7>A} [Jdz)ET*ʪ[d7LJլV3 dT&xC~T~HJ>Gȵc1AGf_Z8/wCkv5.>P@re2Um{5͑߈3p< ? wj8A Hfeo:lo7|( @mf_ zv侾PP5 PTA3ؒE֖\eZ`߉Yt)ޓ5t`k&Z0*K?@gF_QShIL%%zUǓ6ϗ 4!C.} zLjp5yH4:'>赲NqџTOaq:N+&@`#Cb1vѣ_p e,s#Ă\Ec E EB1njQ@=?^';<.'AWR9uH3Zo %ɹV1]#:D!MGON}}pic tӯk 䮏rbsdL:BI/[^V(*i-"YikSB<%6.HQ5*yʥg\(jb׈E:ꤓ_~yyOzB?d^\hո )Kt+Q{d]Nއ)>"Svr1S ~O ZGѴk+'oʆ) 7l+7`%_yfSGWꑷais# _j-za89Z5D\}wUlF;y PNuW֝X%=9)=+~  ,~XD8X8MN 7]qxт%7Eju|2OU۷Cf{h0(&i1 - MN*6id|D}lMq4VA {&p+rܩ| ́Wէ8ed +fLq5T m:<5V#( ̗H2;U^ߜ<"9S*P*Tts68uJNL!fĶC=KH/gG1UFJ,Ze.<*G4#'+Q¬`CŌFv0t%:eCGsCDzov?"vA+@`)J㙖6;&=2iӏ0FUZD<yMuSZ``Bh2=! &ݤ,}@.\’S&H$|J# KJDdI Y 2,XPy`BQqW^8!֫1%G&GF"iRIц?! #Dn${Z`#UNQPצ"R |hyS:tW)%!|55ۘ*fݱƶ0!c y_Z]z:^|k`n>e6>:+;yԨ P%|%*Ƿ'i`7-Iٹ)jְ YP V)`^ aHY!i 9WAe 8%oPmQ4l/Q3=yBBedu lzV{΅P??Cd@nד>ۅ3#j :M%[N{a![!zf.<Йmi2HHr.o$oGKfVgl!>g˕yK**7)Rl+chuN&/JŸ!Tڼd5S; lQ 譨&5;LdKLʲ~(h}{վ4!گv2=L? !RB!p!v%2Y2:^px&!6A2.Q]o# (L Lf(b^W>5p行jâ F\P<#-2HduZ1Eb\#,$"F ZIхH d=2[nNrBlw6'bd eq> 30zЊOyvYLfЅs<k8{.m(n203v]c i9*(%oo{2—GE 1" Bs޺ng2w.'늢 s2zf2C +'3z]h͠FʻU ?TmWP#ۮA׺k6};vjzC{FJvNs@;d}.6w|=iauL3*0՚w[Gf8X'[/- dp\姻ϯnUƖ1ꯄE;"ϑ9u_[L '/z?aZ2@8}[L9/zzyp:|I:L4!ּ$Rt_M=su]gQ$I!Z5˟HoiE=m襻=wؙu"i(٬^\&#f Aj&uK5!J"aa9+jΣn0x|!Rj_ڑ@ֻ=f؍VژڙшdꍯS2jIIIq :02 y(Sj.QnpvWwn`v>zngܦMn|H:Jm4#Oh ?rV$NyL_}8I| ߝ$9Gq;'soFK.<^CKR  B//- 'XZdp9^\2^KdӂI<^KH.4|x;\(L ,EkPA8q]=_=*ÇŔEsgI/OU{w6>Sirv&ޓq$Sdޒ[WNwn2OٙvXR7.]dy#&!h4(hEHvݺj>$䍋L&yr>D]Y-X~nuU?)i7-O "{JTjozf68JjznUcjUšFXmC#6F΃'khFQրP"+PF`N[wtDzy:H$32ԅ|>T8EPDNx*RxQ%%×Pm3>OCrY˸c۵ )_ F,i,UG4)jY+]MAkfFi`)CDuSb>kY{'+; B651V΋ #Yce&3GK Ϊѯr?N_''"{)Kwn?Ov&s<AϚO=N2SjAG <ك-H4f"wIOjKOV$ׄ RDSN XrS!Q:)s=EZ:J4kV\ѷ]]GP˥Bգ~i  imCJrh 5R#(/QJeaߜAHSP|nt 6[scUںLN⫣1NF)/ jb&}utB$Pcv{}S}7㤚]}]įRHVDgCmc jR Ԣj1TI1Ss+^2+;*&}~o.?nG%~7;6L/YB&Fg z6nDr1Z:VKg# OߢpӼ'ܣ%P-MIVVFg hj][o[9+F^%E>;O WGYv$9q#Ŗ٦[A;9T}b]Ȫfl :T E@%ZkP˕=O$~>NϺvPK . юb~Ow0_>Эl^7u9AMi: 3ʓvpAPJHѰ[>[N}\n.U+iyst9a/9!#bwN47/qFdlⰶX붚+Y&ZjHW-ɒޞD!O_MEZ_W"S4)5uk@5UG*ȝJ>@ʌ46cr@IRa,׾^Gh# /iL ƔU#.4pU,$&Bc(HlvmTŨ:eTwb>|[I5%t2xI\Z9Ku6bQt{'7K,7fտgl7"֕?89?9)|_/TE#8O~ ZPr3ŷ2;Ɠ5rߏ]al#EL㝯ƪvW9AKÒ1ANoN>6⿴bAIF5J@iٴ&7OX^~{#e*v2k3D=0\[89`{@1 4 |6~p[A<^Ll^>*YkS[럗dQߌZcuQVAc#sp`5迄;ًN-_eI>wH߸$mXtǀ3k13P1oPNQwӋX~bCu@m#v QʞJCLL`OUGx%E-#**%Pq[GtxiPnjkcᗝ)!Y-J,T6)vּ6iRh$ccqhv^'uTH^g4dEKq$29E|<$6XMaj0mT,g6՚k:^RNUJg!ʶIiGm}_CRVtTۦzե/[JIN!wQnj-5/[JQB3A,u),uy-=}RuZ g!QvT+z RmۤT83ѥڶIiGu%Kl̄'x2~>o% ky`*x ZaOA50(6BUK޸ $\ Ek(h%-A)]鳫iYNK)YFH>W*+q+%ݴ[6G\:7͍:i Fvr 8p>޲eꗚUteU|@Qi.4{sTsTsTs.l|ΤY۔eY(J-:+2;]" DHEYړTrPRE茈JٌGS /B~ &Z Oqp28R ל]h̘tvWX.hµVX*,WI'k;bCWӢ5 ˎ4 OcEán\NPhMu{<1`dg kػkPímRaٴ77^wTX^&۬w]F&D7%!ҏVaU[ȴu8 1p&6WX.^UXUeíK2zkFz|ŲO^̩G+:Cf>O]P^Vtv<4+Ch$>h v!sX $ZtQ̋sLmd.*ÂxCB~²e iCTۦZrsodcx$ŘŘ)O҇Q0&.c0HD1}QaHȨX(LIYNJY.P7F*O_+09RIYr U2$HII{`Ү˂`&B:Ȱif4!h ԪX`7\#@pk̩y{5RHKHsid!ĸT1pmc$Ϯg5?O'>wwȫ)< ́3H] }7Sj}=;?|EմzU= 7[x#H< _T1M`R: -jl#mxo49!)9z`~k^Ʉ٢M F]&:b(! ک0#d)frȒ!Pa8hNӑt"w8dhμ @&"8oݼi'5@ZO^xś.`{kyLnxr^Ni7ʾ.x_{w6˲ Bf,R D^C>8Hwd(jFGͧg^Fz5a7@q#@D&g?Q4ir ,!N.h[(b=c Y^qNI"U~g)|u\!PXs6v5,`i %{ޤçlvd 7eHIKߑO5$=͕`Q?!pmCՊb=Z=ݳ|7rD-EbBAΑ3%X+7~aSގRr?yߞ4_.шQ%]"KgGF;3h^d0~U$X% |/2R_s}rjd=a5|Z51Ɉ\(hA(a/a•^5xg{ 4ʈqn&b)kd-[\$sdf9b5ݸ0Np]#"CKUY39)C1h9)2 'Dh8} Dq2%4cBBBdkS2,"M$/(F9zi7mBodӴZU>jQ ;h!UNޜhoMFC`EiP_G*p E̊O.HLI7cݤT\];Bc܀{i.|z? \ mCf~htDrk1nD%@9{4S O'y!wsC4rC jGNܘ${"H]\t% FL٘\v13s)Z&v]ӛ\[3׷dJiX#4hi <_]ɖ|*O0-l~i }Vߤ9,ƋdFNZ17<{ S7Gwu U[{qq:,H"pBTbI# |oSiGLBIL)s>"9NoX[j5VT tnal!f;os!-w'z(8R(SKq&w ۉ&_۴@ԼY$@ŅzZSWi+~EHicL\r͹f@|2 ~jo5y5A35NPژb  ؁ǯE'䕬Lh@yB^?\4H0lgvMC$hdK@sք, T, HoپNVaZUtPLD>ಡatAKM$N^BThǹ5gf.Q)D`=zDhY<(j /-9yP,%A;ZsrmOJ\S-VB™1E_iˮOys7tf6{}s+ϸ ~0ɻ< [{a5;Bk \;J|(!IS"*ɕWF_EAhr^hcRr͹Ąy!z8~rΐsd%T z00vnf\Imf,sÅXGk,Ϣm: XXl-O˅LןRYVbyd 4V3epL2 ˳z7>'Z+•cGR޶/6CǷv㞭[X\wz7E:?._?.0rPO-aKiTWV|w~b5WQѲ 7ˢwIgw~\ڙ:ڀ,]Tެo}&&ol*` ŖVGǛO͛='jw4]yf++Ru_GW~,?"9[>/6Q m>Vmy,z~-4o-}mn!"ÖU5i 6;fr68 KN'~{[7H:+ mРJre0xH;.1Rkt.$}䱥}:웽ߏ`׎ &رV[e[eG+-m~+:=j#9{z({7gЊlż6 KҴ`klf7wWCk[g̀xvOԶ;#/}WolDdb:ޏlK/gՋںke/ouoQ:w}i0r']O[Zf7)d:qhݎ}*8XvNToE;T &ԧݺuMv;,Bz٭;mkevCq}Sr^ri}ɭ;p\۳ſge%4IsW^<|u+欐['AkoG֯T/ W~#-KtA@{N.* @*Kq5FR YQflHBBGV}`H6QRy8Iy=![X !ͩ_G8Ro9jCm?tyAsZhVP@svʲ3&sMP a`#.Qa58C% *R~o8?($2$g#ےs<)tj`g!&tpsJs-D' }VY96? |u2#H3„cJMepKwcX5Qx㜽q,A2ĕt`es, WA~V(1I)e=]{9Sw7$̦;Ŝ2%L~ݣ}>Zxf 1OX-'+߻t>c\z4NKn}ij~g'lvy1^0$4xz0^|_ycu1 g(YxcXdLjE`IcGJɋ7jG>h4Ǯ'$/ZQI^B?RL^cDJ"͉8M `'TMtYcۆ[aղ yv&_ZP3]fls S/O>7o zH)7|ATPW] Rscd`Miڭ@ANon= 0iiڀ5Jee/!m}P@}N.!Y&X^>GB<%_N2=)mx$В tcQYQߩjlZ.Cj ņU.'GkbkaaJ\6 `\dNڦȩؽ|]/fR51i?U*؉E)DpnD8iTcFX>Hu: 9MTBz%dӨ^X'GG*|h~x9h.$"id x?uV}EK1xgb$P(zq`9]rk 2I'96 K'Ov@0h_[T:.ϪzLJUYARȱPl$%R!YǙmڪ@D3,noH6*/ `~17c#f#=NA2fL1lwJ £ɟVf-Q}<#"hD|5haP)<gL.ncMk733A> MxI,!P\_sm859{__/Ox&uUnXMb)X_T}d9^HLZYp`iCwJ9ba k&:ߍu<Uʌ] T`궞*}_r鰯}FJ lE?&#'ǚ:JE8 Y)OMZ#=dfr;1cv9 =\x!8!|)$;R5/hZʐ{e,Rvf|8T _Kn&BQ߯|9/ДpʙR=1B&@P/2Pm~P3 W*He$FӲ"~ =`""|H+ v*/QL <iA`Xlچ$Y 2>R󄍇YZ \syUE̫e񣝆qܘlRH|ę5cY1їK,q;7a>^>ؒ!K $ċ ;f2IaC%cYE&3*NPgޕ$bm1HV`0 ,vݘy$IhϟȢDR2%eEɞ [/228ƭc`٣yP[)!=2iM^1JrWT6pe `8㞏p na1o0kw"o|fOSQ`V)ݧ&˕.pOMYq'DF5kZYPQPl97W WE15pH}k8i5 /QRS4 ?Pj#gG[C9a0T5fD_6Rm@qAdX\su[hHT?Z NŦOQ/dF@} (l&)^dRJsU AFϠ,q9?`:((-"LHI@L$ YdeX({/ Rq8Yg_O|k09fVٕb .ڈvr.NGmrPl˸SnٴK,.}$`$DD ol#ƿr$0 s$2AfkH>+BkH˂v:=Џfw#2M2U`8y|&a|\ nL6J dyo,qB$,{c1@% Ma510/[NyOPpBR#z5707b\[dIc ]Lhd"$ t9# sI!eQi`_mG\h&Ob+7#I5gPqJF@FBK(X C^|iB\E &Q@Wf711b |n*\eb4&4᫏O 6‘f]-2Ҏ8VGo͏^$T~2j)tרJ\d ZHt1jܖ>lȈ3 ]ehFu)ch#y[4Mm)ihٛ\|}]Z5DU#o8xEC?Lw/qPb E>EϠ\*ʟ(y eb[_c{){ß4rǘE-WH*@/0*?\m(*-ɱgih`6;WdʅXylو)>3/(V5Ys4Z 7_@V@AGT^f'N=&eŮ =$e Zw&,Z@(EZCLOp/ 0dc<|fa|`l]WdVEWVX~8neX${ZyOu9Ij,4XJiRe>/{p! PF;:| `cOe6$$NL'QؐCԙ\RЂ-!,]S)W642RXZ0``xJx*ෟEk%ZnRLX9sAz+JJ21Yc4 _NZbr urJL wR/Q{ ]H8PJPF? ՕPcB/Oϸ# fpz'&J3DFFqkj Vろ1L>Qwݹ%֥J \vivm.|T8 ;3i~pFk?\)|c`*9]<)a,a֮dC.l%HeY+ )i|},9ZFK}X~Lif[PFz#KqM]{yMaJY ˴fdZ(o.%1dNFWL&Iԫ~ oۜd]7j 'wT-Ar>NcٓW>+B꫸U[QN8"h~{Ab|?qV|cnW'0̮_?909_>Cs*o2QS!HCI@)Q ݧN ޏR;Z8~>6OY O*>>:Yk`2'%qi7~~ʚ4B0tp$MHh!(CT zb"tkxRl`xReB(M91p7纙ՌD73QI@T*@9x~>hʩFKqOO`\r/rl/~*L*|_crkmk&OԱ}(fVDVZEr$\-WIE8/[pi)W"#sK2력45d"Tnt%l`~} +44Mb1L(SɥC!tŵlTT&i]k3hEJ@pMi1BC>Rx*s@tixBiyB)rF[˭%]Ȣ$! TjDfﵛWVJ^#ɔ8Z-\rdp9Sq ġ3&HtIķGW=_AV|ۅXÉٔN*F'ALf>\Mfcq)aR^<ڙhM$Qx,J_JST_j,a NHZDM\/!EhX)ASnJŞˑ%b 2JPHZ+J7܏ҳbD%C׽y>Iz/3` XR L=z}-o gxRk9^LZ$' V-Eiw$DϴMZ GWϭqԁDR,Pxމ5CF%aL^L}xQ2`RJrhSl85F5/e;gɄf@+/H T@s*J)G܊Mg7Sr1[e&6X+|@sRrE %Кya={ZuP7T١-uGIQ9Y1:z=m,JPY )JPi45k/Vy<`yn b/?R2{&M(B`-ʓhIc>̮WM F7!/EyAgzH_!2$gd?,<2itOlSuc}#IJ*RTUQWmɪ8򈈌%PH\g:1QzM._B>%_bǬT~Vww G 6y u>\1\1&C4?g@Hsϻ/1nH^IS*h RN(Y{=l5cj޲kl 9p7J-שm?t#Pٻ17>LO?]zTiwV2KlzS[jkVmo[j+Vojz;:ґagЕ`o %n6#+-#)mt͛ЛW$y΢ί鈧R-KqI (&ƩD﷽R#J6}(p %9VK Jj Y2q˦,b~1(V 븵 P!T]~\vȞ6E7^vjꭗBI _vVlP)Z @:=Nx{L Jx)֥sh{tbU t[kF#eGraJ WK"cͲɁ&4 A5TZM.otFlM򧑖H Te ZYFҲp˟&uF+C^j@kYu}:%8A&y\iwf={=FJ9kvm +xߢuS~ߠ j_pmUꨃoԶY3Q31Do.p>}"`0z]Ei& Ss[A`x&#`|UcTv<ٖV5඾=KAP]y|=/ܫń22!+CQio"i%r"@$`1opyFGu¡yCJ6vHi\`YdaʥA<$ͽcϭL8B FղRR!t ycr癗8228Ӓsw=dAGʬQP_V3!&/dX9D8ZpV38)l>^l r{$ďqO}\M$$Mƃ>Dz$x ",Ī(tFF+w %RX_U,0[F(c aa"(qx)E9h"+ ոb>RNV[QΨn{F.* ,#QܷІxf EhvTn n%)m.fqЇ 4ꆏhY-0>t3 rmC]qǍ/R&qFg痣<qb*Nj,,yW!ڙ^]~=ޱE[ ^Ngf$ F3fqR*⹧YWq*O܎yoURN>~(GL@dQEbJۙY 'W]o7S3ɐ}$ MX*5Fi<^ +A8Ϭy2#3Z?H'[ri)襡 4Jҟ/}*h32J#8٠=>L^qG'cvhDE>\<qڠyGSm/th+ 2@{fm׿~A@{\#T 'P$)ACξd zSuG[/۲IE~XsdΣ61p./H]xґp6\@IE_$M/K|s8C_>zWḷ#yʡ[}S365otew$ og7wGf=*6'nj!Cvu1;+==] ff:byycr`#\7]!\Σxx?ٻo3F `Cۍ"ۦ R+:OtLi=o0 O6>yI/iz{1[tAB5/>u#/ˬT*OO9b ?b1y$dFq™{d8{cwMJ. CJ{#`pO5^>dm$O?+#2f>lb$й2O#:M ])Sa?e* `+v5l욄{'q`2T-M@L_6ڢ6@y#ٱ X~J*WJ &c.`g&I2HFBcmF-(l.gqغK֚m,{ TK򦣗`*7ݼH72|vU 6<~4/)ORr,u23OxjL^:7dd!:R ~5Kг@u{jpp B:Dͩ, pQxD*(Auݮ~hJ}O?"F]}Xt"ԕ·D^_R>>Ծ* Ji&DJbV,2v/a^0rc/s~{`^,OۺrMp_Cp۝f)ۧR@>d3R\A8ʗ.^Iy""Jrn2\T7<6cΚiM<&XdzQ!7?ćҡeyL>F#з`-tEQ&3,ZDícLG&PJw,9‘ Kj|6;IUɾ2c߳upf/!FܵaNK [(˭r$[O0PZwWXB`rP(3U%)VRZD z{e6Qus5YU-gfq7ntX6RZ'L ?iX":6ٶU+PYyXW`F*I?}5Cd't'ǀ` HCڇpKh=h`PaR24ƍ~_;OPz2΅&ks/WW86ԇ .T ~>RKoxz9 boL@|ÄNDe2`e&$oO݅Y7"\~p~eTɛFg/'^,9z!4 BE\fXIQ3xh [J%T; O;Ttg,!*NG~W-ƦNmhF)xnVw|񊌡X+*;)A}GGQ *kp"QE] |iO;䘧`8{#r=IzQQ(yT7-eC=2jR 9nzVY\3HUF爳F4RB!l[uWJSg,|wtL8WPp>kt1H8' BY3N[`^c#)M>xfi/ϞPN^r >\Nf<8?Q>=9VИ')0MCO?fX]|4uAQmN9Ց22#[Nfme0b2"yyYRN4>5(Ofe@4Mu2p'æƻrQr/ov$ҽ=pDtG{P OJiJ.~v ɜUyv؉Rv2%|ιVSҜY+?B#"cHa>;Gsw?:fm~ۨO/?6zPoW Khu=Q@l:^^vzxkw4&57GRzްY@3id'5 f F YP' Zr|%ܦ>_'J[7Ng[B )GQm)IB>gac?$<*:eNs&Я?IA%Q#2M6 ODeT}_3 7rz>8pOk5ij[a?7ifxvf2WsZ}ݤ_F:'gN-U^ ~6l茒2y՘jL^5&_J4("%;L& FUlTbʡl9뺓~hT؏SZ(a)F{|OX~v$t Aoݾ<&Gl{}D"c)Jh_ZP<M%%)Z  H`U0ϖ}r!C~ZO Ajh+.}pfɵ:&9qucܲS܇h: 'giLxіbY$հ@NOv>^8Hs(ӘS{%zhtU' oig2Wf?W9W8O3( rv' e.뻉N|ӓQؔW4]Bqzb1FsJ0y˫+ު/>]۾\Ң(l@"@I"TgMyqdz6:yx /i^H?utj: ^GoB2X[N-U^C_y Po-('% MǠtqA8#Ndk+Cki0Lr#Uc{.9y$YxӤtQYrRid(Ӯ FC,rrJ^R%Dr6GiP%BAJ0XUYXJ>_B!4,lo;sU]03vv!IЯ~tWFxNj :O7F]vw&GHPtTnuZH}'I+ Zc3vu,ew'_1UI?GbslTٖwȫbG[cfs|V\gEè&UŪ{ް7VZBrOawﻏ0@N6]dzz͔끾 QҁJّםU0]5w{a:wX=}z3?L.lE1zV7e3i]p:[H&;ޘ!77';ݧ&XĠ/.~Ҁ.RSQ&-㐒eāo|=x[=r[OՎ7iiPQOnx+9a0&uNx7Xֳ&B!,I,w%R)hT6JL.`)x}i7^6ImUo..a ~݊=ۻf/v}wCKV1K^盍0EbK@ƬH'+3U&# V&'~z\G<"sBJEH?Is|Wg$֗wM^{5_?YRT±56x35_ZvXgx j%[T=9*[ic^䋀Tp^%\]ia 7 TvNW"뛿xz7Õy=9x]Tv^VȒIZ+є$LKT.7Wl UeiLhw~E`*ֈ@oZ6ǢcVYѱ o.ˣN^^ ݅H7e,fywFki.4 /sݘ; Wi[LD'Zac\n8(LM'&|0ţ,kgo7R4HFe dEY 'TObӖWmseX@YIR[pBĄPPP"B1nٰq`sq,m?x3`Lys)]htu[uokew_ޮW6jZ* c\kyX;n.Ju+1BHYKx3?&AwWVSM 5Pe'J,",$ʛ ~_y s,<s5@Lm*-lVҸ#01Dp8y "z5oFjAEva%Ib#x7^;zj;M !V *'%ߏWik?xտ.cpNma {)1exy{w?#e:e bcJ w9vR7'fғ#oVb*"PeoJ`PfkąF&h{D-H‘ql){ =o8T{E Wނ(J69oED/bo-E9.7"&ȶ&/.ެz 0&yAG<@?8G=F峧+2DmOY.^:iс3rIѱF Kv!E?&E{J0PY2,Q;i[:%\m"I-ФN`e"Qx2IxUw`|,Pr[7Y˺r6ԘYZ4%*5GC󇎯UgR t1d3}Mx妮jxZU^dv ͆v I)!S1a 5˼+ ^K_¶$6clg<[<I+L-F{M˽c~ڏ3ʭ 42fBv*@ %Qbv"h "ZzRX;ZO(-W~0"m[X]]J\KcDWU,!F~݇x0 Z@:'OŅ)Za,ԖI1A-G@D%#!iɮUSNb31x9U3N@=֗V졵;ٵ_Ѫ2xJzl~\g.T'TX.5h. =ɮE'kF(hBdf=z }/AY`%$p*G!:`47?7|~s{?^VOWEYѴ*$rwXl xx*fe}l1Į8lCp@ Hz}GhЉw.ӛ<ZauV&ɠٵV" 2JFm%|XjӂskoZBEmO h5V3h XIu r/|u|O#q$ixױE` +AY8Y )3ֻY`58 3n+yֶFφh)s.yיʠp!}8;jļevy$qEi$+Zi @U l-T ,_SX|x/:zG :82dXS|yQ\ƲԴ Z؂P2ϘbTq>הnޓqdW}Yǫˀ>%` ̧,uJ$n}EhIT7eڱDEV;]D}\hq{:=|hfCA9dHַVk[wY@8sba"^ 4\F%p0iΘ1Z{4.`[f1= ;gSf6V96@J:e9 6F.PrW0cܰz9i]亐fQy=:Ҩi p┋"I(_+N8gJ|5jf sq ȩ^tiP BT@1g  =>;_uCI0:$_x,TIQDZMӤidEC(Q2k Aq֥lO*J$h#@ySm A8(6}+^}Q]4N[8_Lkwm:eق$暘5I)HS,#BT&n&e U'7~kv^OvA"v?'^4A֏l- 7f8=(zB1[΢{6jZBfJQ[G{B+]uwj5c8X/] SÁl*kA(z M I \{㋕>6w0OnSI*>\v=6E7^FvZ+f$}O~YY/"oT:JG!Ҭ0՛dJ}<_g<?&|#*+K8h,qI3h/DcBIfnr|E ڞ_$TЊcA%c^5oòbճY@W ), ,*64ko KJH#Y<ڵhY@-ԪA>K֨t1J[JmZnET=yޱz4t Nw9g: 0hG$o7uKQ.>8G4DIaJGio|'%i{۬Yt5:tY{4=vdAoסb]$(z&Γտ qț~1W›=[g| Bբ|i#m@ '|8-i  ЃCߣ޿Wg`w:&w]݉&hw8N:lt. ehdz\hXirS-KԉKaXZH1GmNPS:oG-[]phY&*(҇rRxH:-iN,VE…hBP1sOزa&0|O| ˠD(L8chБ2uԃRTb"UbRʖD'P K(Ye@<*R{U.h6_57y3JB)J[u3 y)5h!^d}4[w&DOEVe[C rh@s[FC^.,c²[CTnKF]J%"(%3;Q.+Ա1HAHYkKy' "Jh] z rr6*z˽Մ[Zj/*2 !4Pv[1QSpb+Z{E#N%ж&L,#о`Zݦ N3~zrT|kz ̙8:IJt߼f}4~ol? Bo лF seUFa >̨T&2+TwG2$E QFb רc&X|!Ӭ}Eda#j$T>,{z;2{zA"P_'@ٞy RQsz਷lzH}=V<9ɗsJ4?9rS2^ҵG!`¹hPB?W A\s" &s>J}.V z&|a~V@y*]hb(#<7'}6^X8rHD_Ɠ*RthΠ/?nJGh99M|q`TWP'Ǔ4C5ڠk*{By>#իL;~|=%g˖)[ 5pb|;Բ{CKf?ǫ g~j8IcƯVxû<LߍFck@Yyae"C,*'e`e+;@ B# 5 JSە^Y7;DFFߨBEtftA^5TV'W8l??/^?Y kܟ̶Tk_b|pdc嬯BFN㾥[ӖҭhLD Ct늃6H/Ehhoֽ^KYO,w0 ]aWrNv<=`ZNWK惃rwcjT3tG7^Yat́jxWqp}TdM) @@ AJ>qRbeDNH, ʖdeтyVRUJ."icN (y|BC7!1G`ԚP3^z7i|J:s6[Ȓeb\%Ge2wIzH?hkFr6؝a6xա{1])/E/fq$*GzC~YwnW-Q4G,`9Y5VGU!>ZܸfMb-2GoA&-;ڊsCj)qrٸ9dv/Ԩ&x`fXWYےn >B"T9dN*{dHe! 3/ͮFHGVIҡ(ڻ@?u |W6Y;9Rf7A^~* 1( FO0,2FIKO'3)$e߃+J~(=KIQHJs8 S[j6RF :dm+ T2ݬhe/g+]mo#7+|ٻEZKH{8v/`o$0:vF;<"ȖldYf[Vk,L0RX,0+oZ$n}_A4X<@3'9`x/D\\iXw9 *gâ2A*"X Z%S2, yj7c`Rnth`,h?h: *xڼjAz.|2goJXKlTPڜ5WRD\."HkLN\ IUы?Ǔnjq_4^r@ߜ|sj.:+;&Xa8:̑r]/|f!LO"|w2Ma~s2pu{~zzӻzrꃕ^n?Om&fz=b]#ωD\e_ or xXw{0}µGּHg_N~!RF$_t(|DMLM=3Nc+( HgIz;4 `y )r;; 'TTG}3RhY/L/[߿RWM妓*}$b:AOV,@j~!ۻfyXYY8K^Y:Y r55tj;/'ӡL3uݜ7&E5`NJi ۃRڳF%b*YI%BC% DKS+\<*fz@ܠn(IOB75'~>.#&n28|2?w7s;/ Z%iEuMx'!;%z~ss3X6$(]2MFq}ߏ;.̐Q<kdž_:/=e6f*^Ԥ|pE3n |H 幷tp;g$<牽 QWUk :NGJ4kUN vnz fx%l6`ZO\m+tNr,0W4FB%Nx4"V: V)Qr#`r//J`uTFMrKF@J*Fӥf1',0JL4ygqnoG{ƶ~EIZ IUu_>؊3BqR4:X^.kǨZp"Jl<ɤeR@U(,6%+m I}q1PjI>+LF;.H:!BA&q%:1% ac R_L'd&Wm(W(!K)w(-~On/|ك0^)%@ Eo.ØF;<{Sxbp~fZ Ga $f,Jha/Y[xg?q.gq@Qq 4h#1ctz .09@E0擻J,+%m"'e##JU&ޱ}:cDZibbo] ƞ\[1_zk#D g7*8䪮OnrYיN5wT)gE D#2fz%TS/?hoA0r+,{ _`>$J?ȏ>\D4$ȒWʧ8(1=b!kXy#60C&eҀMzn4͒'َy1.0z}B*CI=j""JFLe>BuQpv gJ )},98JBX/9jSDSH$ J]ԀmZq{҇>e ۬T *dJK5jN9Z\[J QPݸ @8ч q F@8:,YzX4wvmw`{%Dp00$,' hFRApt?} Mtקx7$ !Tc 2$z1 ;gQ#)hr_ɏiPQ# iiLGiuظv/5n.uиxa>F]9t)N஢Ċ`:dji33Ucf3CV5G_,]d:ヒ/ތv%cfgh7E7ۛ뻏W/ nr,rBe%R,Ue(wYw_[<=iW3MQ!vY!lDJv[ 2V=N&Z -Oc 8`o ,t`Jԕ{dTzYz t4T%1*Qih%ʥFӿi+8î򡳻q!@`|DeV(v6Wͩ7ZឍzV-}#e=r9yz.>ӭpe \Vڒ*퍵 g N۞W-2lOk+DQ1b!oB @*H~z:u 0iն>Xco[7Q}JRUo[pmVJ tTAKG+=l+#F|6X!n]weq{33s7ʍIY7?{UePTFgg<[c5>;<8;'Wak pвq;[ 0l+xVp86 d@j/P}J5GOu&+RJ;RZfsuhheGZ'h׍:}Z&WNcwVZvsVZv^P1hhʬTҷ$8TΩY VxjZP%}–řd޸.OhG}PpI@ϴ_{Pb h- GA0I[+ڎhsv+_!SO<ҏYc܆Hl"aq!+c@ȝ ZU|Njx i6F|6x]PTQG[H0юHF7JLAn6Z1ȏΓ!$s/uv<_G}n<߭I ɤ#D~Yr]\t`SFfH ) 8#p:M`|yr+#lZ$ɝ\Zn C*AbCV:h$p<Vm<(SZYv X_,BfϟPb}Ld  Fu0N$4v1lY`>^4Fb9TTTT*zs1TTMF/PQ4 ] TT+HBHVVBΤ+w]c9SNE!+b(FqL9)NVL# ΘR7LD2EB5WeT\-&Dc>WekZq8% ӸNRAX(LT ֘}óR^xVz/6S}Jpj?p+V*F9mdG\dUbSR+H&H2\Ic2DOhTR,=碈,m@&JqZm+Ne)zNaRTt|GVE{)Yt 0dPT#֤q^xm?f ?l+ blG$C߃ʺTFBrdisjP˕8xi2,7 }5/m˜N\m+;Hʂw4=9ʌ;3ژ(`!f$@ FFKsݜj־e ZIwk:gM%T=ZV(<1Qqٹ;u5{:g&Ԥy2ؙțqs7$p};RI}p-h @eݢͥ@s2z)jQ&XâWLm Z 'pks05o}uC_C&+FVB|OA.l %T+󐁤; LB$@3_Z"ofΟP-ēy֜5+K󬉻~[9Y4/mL[HۓZ0ugws7Jի\]eӻմ}_5g&L뗒әNW oSP'av7"yGz3&߇YCӹc!oDNwSuAt6mռ[{Bևq]»\()gPvgPt[g kcFT3.tZV4Φ=#2F - <2%N-'}Y=$pRF5AܥsR!T#kL;ޥ]:V| +>jEr.G+=l+ՅVGphOSJhVp^̮o"|X ![UlS]]d.8gk d͕'t?>4N?=bۊpۉ~T+dݗdLt^X؃!iGml&3ЋכZ~ 'LqQ'RGcPFRψ#UOϊQT:wlF!vHE|W.pLr7{{^KO5Hjl‰;?Ju 5GV7tsW[C6u3KJeϴY5Z8OOӉRE # C&zJFJR?:)i6f.:g6.=s1?VnTc_T{ӼȦyM"E4pJQ:D⼲-SVDP1PJCz.ΗzZ I1'Xqe]dQi].ɆEi2s\_bК鈥֥.ŢP ^x $ko ۻ&;n_K*:gIAWGoIY͋-GsT}6C66Qֵn$w"*Zpt*21J PB1&!7FvqcT{ `Ibxv.>jUx(PƮH{as-K9Ħд2zDac裀l,Xaf 0yڣ_>0@)6,\~\bкwfhgR:s K}a;V?!;ITaf{_z?"+U˝r`G"y/QmCN`Xv#+ 1"n.|V9ܐV R)_G7J9d> *Ps6ħ) Iyajo?w7WMu_.p߭?͍XK(OR au{ǷWaoxS܇o4I?Ƿ٫䯷;3M좳?~Ool[̉ټ~x}vB u>ݑt42:whn`'SVJ>>i=EMnW//KHg1 C5T~t)E7tc1a<͜ךa≍89; ||dNaeHznS Ӻh&DFb٦+Al$P8ȍMa_ $eWC*nrJ߀f,HK5C#%J)K!J fS@=gs{fs) 3B"naE9d)+LkXqF\J@T`գ>W'˛~l$(8 l=`/ L SJn2#`ًM+lZN3an7fIkK#҄{߼&QiQ y=Bb:C~pos aa/2!V+Cj,ڹG@(t$Kk@7PyDJ$Jݴ urne7)a)8ɭ뛫ve''T|k}`5;0Ąk]㼶8c)0HZ0;O%'VlYɮ:0Ԭ_>2+U$d_/>帺r~ٕfWlL5C&|u,\\!lv&e]ܮ~6/of=|i˧GYR-6ڇ{gy1Fh/(b|Qͮc柑H>(=*;)Ġh0hRpY@qӀpeWR;rf ;S[#H@WDF;8 ͲGoj Y`h(mBhqP:*Fo!AMLu{b!5+1TyȫU?Xj-*IBIH),mT-5ʄz/x}ԊX饷 R_2~{2 o#jEݠ鯍ʦqMV.ɼ)hLC 1TXgc!`khTANwo4PZЫ&|L@>rl4V;l6 GlN6>TCdR1{? * v72 5AfĩZwAaiV)pIJzz)$רDzΨC8Ug;wkA׶l\Bj@nL`1hP!/"Ç`I>aӑJ&U7 7V鐬0X:e/իyƷ+ R c?S~E[.|+ %jiJ-Psu^I0[ OUb-o=Ykil=ZAzbT-zI3 $nʈriתҍ]g:|廾] VWI`"Q | Ypڱ 7 6FojVƬt/qM KXXV7iZ)!u8 NmT Zgk8/K2<$ &'P`~mA5:r|q GTMPS,CtD$ i1srl$/_(Rn#.BpEIG &&Db`5!f&vcƶ<(2A1'r*X#V<4,nF֭MZb˴+ 먅s:2U*-m3۲2 DZʗU/+ aOݗnCmXRW0kJ~ђ)0.tKA@WYb`A ;U܎3'6i'<"JRbnݲitLshi H+vLsk'՗vݿv>].ۉ?l7`@uo]}{|t{?U-՗=1?j_.E.=k3ue)5HyH'R[xT|):3f+$ t2>Kg4;Ohw/plcLmZ|8?oN ެ,MZR\y0 l̛ VE]@/_] =dݹaDp:{j#x>27=_ó1ohJ8pͼyZ<Y  zꇛhSp*RXX*qlRVW ژd/Z&,s( .@.ul EM- Ni>'i y&ZcSJSƴ VA~Ļ-wu:1՗fz1,䙛hMYWw tjQݖKyNG7w\ưgnm 96se^C̕9o8\%3/wtWdQ$ N}UgvO.6Dzs'x㚭q&E +<aa\J6.rwWz}fS_)kDpj)'ܠ^oLB7)n.|T+Y] 2v]AEGtH0Na ?8IJiCdrwi2>~ɱɜ95TQMKC״:;+Ԧ׷Rj! '/#l HR蚅^;QCkj.@5[0J2dThniRk)192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.131227 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:47382->192.168.126.11:10357: read: connection reset by peer" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.131307 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.131499 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.132955 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.132987 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.133004 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.133690 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"2cf827947c686099ca3c6afad51d866f4ee1d557bc64cc1c70f6213fd4198df2"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.133908 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://2cf827947c686099ca3c6afad51d866f4ee1d557bc64cc1c70f6213fd4198df2" gracePeriod=30 Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.200566 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:15Z is after 2026-02-23T05:33:13Z Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.423786 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.424354 4778 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="2cf827947c686099ca3c6afad51d866f4ee1d557bc64cc1c70f6213fd4198df2" exitCode=255 Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.424422 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"2cf827947c686099ca3c6afad51d866f4ee1d557bc64cc1c70f6213fd4198df2"} Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.424463 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cc7b2a12646c299c75286fc95cf2a8fa35fd83031ce3daebec42030d966274ae"} Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.424551 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.425253 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.425278 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:15 crc kubenswrapper[4778]: I0312 13:10:15.425288 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:16 crc kubenswrapper[4778]: E0312 13:10:16.024601 4778 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:16Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c1a09b06f665b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.194251355 +0000 UTC m=+0.642946971,LastTimestamp:2026-03-12 13:09:42.194251355 +0000 UTC m=+0.642946971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:16 crc kubenswrapper[4778]: I0312 13:10:16.201021 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:16Z is after 2026-02-23T05:33:13Z Mar 12 13:10:16 crc kubenswrapper[4778]: I0312 13:10:16.413672 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:16 crc kubenswrapper[4778]: I0312 13:10:16.414799 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:16 crc kubenswrapper[4778]: I0312 13:10:16.414863 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:16 crc kubenswrapper[4778]: I0312 13:10:16.414875 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:16 crc kubenswrapper[4778]: I0312 13:10:16.414900 4778 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 13:10:16 crc kubenswrapper[4778]: E0312 13:10:16.418296 4778 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:16Z is after 2026-02-23T05:33:13Z" node="crc" Mar 12 13:10:16 crc kubenswrapper[4778]: E0312 13:10:16.430298 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:16Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 12 13:10:17 crc kubenswrapper[4778]: I0312 13:10:17.202311 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:17 crc kubenswrapper[4778]: I0312 13:10:17.253304 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:17 crc kubenswrapper[4778]: I0312 13:10:17.254907 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:17 crc kubenswrapper[4778]: I0312 13:10:17.254969 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:17 crc kubenswrapper[4778]: I0312 13:10:17.254989 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:17 crc kubenswrapper[4778]: I0312 13:10:17.255864 4778 scope.go:117] "RemoveContainer" containerID="37e5f37a4b11b4c76513fd0ebf7036a8c7d1c8547248167f22dc4b34cfae47df" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.201645 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.433459 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.435146 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d"} Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.435434 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.436132 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.436284 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.436722 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.436769 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.436786 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.437116 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.437143 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:18 crc kubenswrapper[4778]: I0312 13:10:18.437153 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.202161 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:19 crc kubenswrapper[4778]: W0312 13:10:19.432767 4778 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 12 13:10:19 crc kubenswrapper[4778]: E0312 13:10:19.432845 4778 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.440460 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.441426 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.444914 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d" exitCode=255 Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.444968 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d"} Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.445053 4778 scope.go:117] "RemoveContainer" containerID="37e5f37a4b11b4c76513fd0ebf7036a8c7d1c8547248167f22dc4b34cfae47df" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.445296 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.446711 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.446780 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.446809 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:19 crc kubenswrapper[4778]: I0312 13:10:19.447736 4778 scope.go:117] "RemoveContainer" containerID="d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d" Mar 12 13:10:19 crc kubenswrapper[4778]: E0312 13:10:19.448101 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:10:20 crc kubenswrapper[4778]: I0312 13:10:20.204830 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:20 crc kubenswrapper[4778]: I0312 13:10:20.450575 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 13:10:21 crc kubenswrapper[4778]: I0312 13:10:21.128136 4778 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 13:10:21 crc kubenswrapper[4778]: I0312 13:10:21.148633 4778 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 12 13:10:21 crc kubenswrapper[4778]: I0312 13:10:21.205529 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:22 crc kubenswrapper[4778]: I0312 13:10:22.188904 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:10:22 crc kubenswrapper[4778]: I0312 13:10:22.189262 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:22 crc kubenswrapper[4778]: I0312 13:10:22.191096 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:22 crc kubenswrapper[4778]: I0312 13:10:22.191159 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:22 crc kubenswrapper[4778]: I0312 13:10:22.191197 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:22 crc kubenswrapper[4778]: I0312 13:10:22.204937 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:22 crc kubenswrapper[4778]: E0312 13:10:22.356833 4778 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.203072 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.418745 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.420479 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.420553 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.420576 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.420621 4778 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 13:10:23 crc kubenswrapper[4778]: E0312 13:10:23.425896 4778 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 13:10:23 crc kubenswrapper[4778]: E0312 13:10:23.436900 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.534992 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.535331 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.536912 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.536980 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.536996 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:23 crc kubenswrapper[4778]: I0312 13:10:23.537726 4778 scope.go:117] "RemoveContainer" containerID="d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d" Mar 12 13:10:23 crc kubenswrapper[4778]: E0312 13:10:23.537942 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:10:24 crc kubenswrapper[4778]: I0312 13:10:24.206444 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:25 crc kubenswrapper[4778]: I0312 13:10:25.189907 4778 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 13:10:25 crc kubenswrapper[4778]: I0312 13:10:25.189995 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 13:10:25 crc kubenswrapper[4778]: I0312 13:10:25.203652 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.030717 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b06f665b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.194251355 +0000 UTC m=+0.642946971,LastTimestamp:2026-03-12 13:09:42.194251355 +0000 UTC m=+0.642946971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.035255 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.040096 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.045016 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.050174 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b9f3e256 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.35392879 +0000 UTC m=+0.802624186,LastTimestamp:2026-03-12 13:09:42.35392879 +0000 UTC m=+0.802624186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.057669 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.354787306 +0000 UTC m=+0.803482702,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.064863 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.354803178 +0000 UTC m=+0.803498574,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.071772 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c7f7c1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.354811408 +0000 UTC m=+0.803506804,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.075729 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.355802648 +0000 UTC m=+0.804498044,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.079909 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.35581722 +0000 UTC m=+0.804512616,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.082948 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c7f7c1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.355827061 +0000 UTC m=+0.804522457,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.086985 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.356463415 +0000 UTC m=+0.805158811,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.091357 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.356477246 +0000 UTC m=+0.805172642,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.096412 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c7f7c1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.356484557 +0000 UTC m=+0.805179953,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.101335 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.356803539 +0000 UTC m=+0.805498975,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.106889 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.356832252 +0000 UTC m=+0.805527678,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.112650 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c7f7c1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.356847553 +0000 UTC m=+0.805542979,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.117271 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.356957044 +0000 UTC m=+0.805652450,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.122045 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.35701414 +0000 UTC m=+0.805709546,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.126846 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c7f7c1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.357024771 +0000 UTC m=+0.805720177,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.131160 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.35850399 +0000 UTC m=+0.807199396,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.135870 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.358516441 +0000 UTC m=+0.807211837,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.139985 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c7f7c1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c7f7c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250387393 +0000 UTC m=+0.699082799,LastTimestamp:2026-03-12 13:09:42.358525592 +0000 UTC m=+0.807220988,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.144333 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c5643c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c5643c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250218556 +0000 UTC m=+0.698913972,LastTimestamp:2026-03-12 13:09:42.359345795 +0000 UTC m=+0.808041191,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.149241 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c1a09b3c6c56f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c1a09b3c6c56f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.250308975 +0000 UTC m=+0.699004391,LastTimestamp:2026-03-12 13:09:42.359356106 +0000 UTC m=+0.808051502,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.155177 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a09d2ec8392 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.772876178 +0000 UTC m=+1.221571574,LastTimestamp:2026-03-12 13:09:42.772876178 +0000 UTC m=+1.221571574,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.162222 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a09d2f1f29c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.773232284 +0000 UTC m=+1.221927720,LastTimestamp:2026-03-12 13:09:42.773232284 +0000 UTC m=+1.221927720,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.169711 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a09d342598a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.778501514 +0000 UTC m=+1.227196930,LastTimestamp:2026-03-12 13:09:42.778501514 +0000 UTC m=+1.227196930,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.174588 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c1a09d3b75462 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.786167906 +0000 UTC m=+1.234863302,LastTimestamp:2026-03-12 13:09:42.786167906 +0000 UTC m=+1.234863302,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.179117 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a09d4bf7a2f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:42.803479087 +0000 UTC m=+1.252174493,LastTimestamp:2026-03-12 13:09:42.803479087 +0000 UTC m=+1.252174493,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.187138 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a057ac0f1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.621058801 +0000 UTC m=+2.069754187,LastTimestamp:2026-03-12 13:09:43.621058801 +0000 UTC m=+2.069754187,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.193437 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a05913be5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.622532069 +0000 UTC m=+2.071227475,LastTimestamp:2026-03-12 13:09:43.622532069 +0000 UTC m=+2.071227475,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: I0312 13:10:26.198255 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.198336 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c1a0a05c87177 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.626150263 +0000 UTC m=+2.074845669,LastTimestamp:2026-03-12 13:09:43.626150263 +0000 UTC m=+2.074845669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.201529 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a05dcd64e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.627486798 +0000 UTC m=+2.076182204,LastTimestamp:2026-03-12 13:09:43.627486798 +0000 UTC m=+2.076182204,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.205304 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a0612d4ca openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.631025354 +0000 UTC m=+2.079720760,LastTimestamp:2026-03-12 13:09:43.631025354 +0000 UTC m=+2.079720760,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.209792 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a0647c085 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.634493573 +0000 UTC m=+2.083188979,LastTimestamp:2026-03-12 13:09:43.634493573 +0000 UTC m=+2.083188979,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.213856 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a06949447 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.639528519 +0000 UTC m=+2.088223915,LastTimestamp:2026-03-12 13:09:43.639528519 +0000 UTC m=+2.088223915,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.218419 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c1a0a06a91d05 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.640874245 +0000 UTC m=+2.089569651,LastTimestamp:2026-03-12 13:09:43.640874245 +0000 UTC m=+2.089569651,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.223089 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a06ae8b5a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.64123017 +0000 UTC m=+2.089925586,LastTimestamp:2026-03-12 13:09:43.64123017 +0000 UTC m=+2.089925586,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.228388 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a06b97ba3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.641947043 +0000 UTC m=+2.090642449,LastTimestamp:2026-03-12 13:09:43.641947043 +0000 UTC m=+2.090642449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.233324 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a07f1b8e9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.662409961 +0000 UTC m=+2.111105377,LastTimestamp:2026-03-12 13:09:43.662409961 +0000 UTC m=+2.111105377,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.238380 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a196f74ff openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.955862783 +0000 UTC m=+2.404558219,LastTimestamp:2026-03-12 13:09:43.955862783 +0000 UTC m=+2.404558219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.242057 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a1af68cfa openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.981493498 +0000 UTC m=+2.430188924,LastTimestamp:2026-03-12 13:09:43.981493498 +0000 UTC m=+2.430188924,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.248936 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a1b100281 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.983161985 +0000 UTC m=+2.431857381,LastTimestamp:2026-03-12 13:09:43.983161985 +0000 UTC m=+2.431857381,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.256136 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a2cb2a380 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.279032704 +0000 UTC m=+2.727728140,LastTimestamp:2026-03-12 13:09:44.279032704 +0000 UTC m=+2.727728140,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.260565 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a2cc23d39 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.280055097 +0000 UTC m=+2.728750533,LastTimestamp:2026-03-12 13:09:44.280055097 +0000 UTC m=+2.728750533,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.266240 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a2cdef7ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.281937835 +0000 UTC m=+2.730633261,LastTimestamp:2026-03-12 13:09:44.281937835 +0000 UTC m=+2.730633261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.270664 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c1a0a2cf61c46 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.283454534 +0000 UTC m=+2.732149950,LastTimestamp:2026-03-12 13:09:44.283454534 +0000 UTC m=+2.732149950,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.275705 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a2d4e1965 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.289220965 +0000 UTC m=+2.737916371,LastTimestamp:2026-03-12 13:09:44.289220965 +0000 UTC m=+2.737916371,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.281785 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a2e8c3641 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.310068801 +0000 UTC m=+2.758764207,LastTimestamp:2026-03-12 13:09:44.310068801 +0000 UTC m=+2.758764207,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.283479 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a2ea66ed5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.311787221 +0000 UTC m=+2.760482627,LastTimestamp:2026-03-12 13:09:44.311787221 +0000 UTC m=+2.760482627,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.288258 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a3a407626 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.506431014 +0000 UTC m=+2.955126410,LastTimestamp:2026-03-12 13:09:44.506431014 +0000 UTC m=+2.955126410,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.292910 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a3a495e9e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.507014814 +0000 UTC m=+2.955710210,LastTimestamp:2026-03-12 13:09:44.507014814 +0000 UTC m=+2.955710210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.297589 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c1a0a3a4b53bd openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.507143101 +0000 UTC m=+2.955838497,LastTimestamp:2026-03-12 13:09:44.507143101 +0000 UTC m=+2.955838497,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.302343 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a3a506350 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.507474768 +0000 UTC m=+2.956170164,LastTimestamp:2026-03-12 13:09:44.507474768 +0000 UTC m=+2.956170164,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.306942 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a3a53868f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.507680399 +0000 UTC m=+2.956375795,LastTimestamp:2026-03-12 13:09:44.507680399 +0000 UTC m=+2.956375795,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.312042 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a3b88d1af openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.527950255 +0000 UTC m=+2.976645651,LastTimestamp:2026-03-12 13:09:44.527950255 +0000 UTC m=+2.976645651,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.315664 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a3b99a073 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.529051763 +0000 UTC m=+2.977747159,LastTimestamp:2026-03-12 13:09:44.529051763 +0000 UTC m=+2.977747159,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.321900 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c1a0a3c134a22 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.537025058 +0000 UTC m=+2.985720454,LastTimestamp:2026-03-12 13:09:44.537025058 +0000 UTC m=+2.985720454,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.334978 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a3c145eca openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.537095882 +0000 UTC m=+2.985791278,LastTimestamp:2026-03-12 13:09:44.537095882 +0000 UTC m=+2.985791278,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.343407 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a3c160604 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.537204228 +0000 UTC m=+2.985899624,LastTimestamp:2026-03-12 13:09:44.537204228 +0000 UTC m=+2.985899624,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.351665 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a3c160f1e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.537206558 +0000 UTC m=+2.985901954,LastTimestamp:2026-03-12 13:09:44.537206558 +0000 UTC m=+2.985901954,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.357473 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a3c2d3201 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.538722817 +0000 UTC m=+2.987418233,LastTimestamp:2026-03-12 13:09:44.538722817 +0000 UTC m=+2.987418233,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.364314 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a4592aaf6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.696367862 +0000 UTC m=+3.145063258,LastTimestamp:2026-03-12 13:09:44.696367862 +0000 UTC m=+3.145063258,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.373092 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a45c620c6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.699740358 +0000 UTC m=+3.148435754,LastTimestamp:2026-03-12 13:09:44.699740358 +0000 UTC m=+3.148435754,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.378642 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a467e4df3 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.711810547 +0000 UTC m=+3.160505953,LastTimestamp:2026-03-12 13:09:44.711810547 +0000 UTC m=+3.160505953,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.384134 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a468c0832 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.712710194 +0000 UTC m=+3.161405590,LastTimestamp:2026-03-12 13:09:44.712710194 +0000 UTC m=+3.161405590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.389966 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a4709016c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.72090046 +0000 UTC m=+3.169595856,LastTimestamp:2026-03-12 13:09:44.72090046 +0000 UTC m=+3.169595856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.395767 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a471795d8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.72185596 +0000 UTC m=+3.170551356,LastTimestamp:2026-03-12 13:09:44.72185596 +0000 UTC m=+3.170551356,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.401634 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a51622f10 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.894517008 +0000 UTC m=+3.343212404,LastTimestamp:2026-03-12 13:09:44.894517008 +0000 UTC m=+3.343212404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.406576 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a516601a1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.894767521 +0000 UTC m=+3.343462917,LastTimestamp:2026-03-12 13:09:44.894767521 +0000 UTC m=+3.343462917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.411292 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c1a0a5256eec0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.910556864 +0000 UTC m=+3.359252250,LastTimestamp:2026-03-12 13:09:44.910556864 +0000 UTC m=+3.359252250,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.415658 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a526bc931 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.911923505 +0000 UTC m=+3.360618891,LastTimestamp:2026-03-12 13:09:44.911923505 +0000 UTC m=+3.360618891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.420171 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a527b131b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:44.912925467 +0000 UTC m=+3.361620863,LastTimestamp:2026-03-12 13:09:44.912925467 +0000 UTC m=+3.361620863,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.426989 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a5fc2b160 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.135722848 +0000 UTC m=+3.584418244,LastTimestamp:2026-03-12 13:09:45.135722848 +0000 UTC m=+3.584418244,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.432716 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a608f1168 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.149116776 +0000 UTC m=+3.597812172,LastTimestamp:2026-03-12 13:09:45.149116776 +0000 UTC m=+3.597812172,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.439609 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a60a53c9f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.150569631 +0000 UTC m=+3.599265027,LastTimestamp:2026-03-12 13:09:45.150569631 +0000 UTC m=+3.599265027,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.444812 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a6a31e74e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.31078331 +0000 UTC m=+3.759478706,LastTimestamp:2026-03-12 13:09:45.31078331 +0000 UTC m=+3.759478706,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.451525 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a6beee97e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.33994739 +0000 UTC m=+3.788642786,LastTimestamp:2026-03-12 13:09:45.33994739 +0000 UTC m=+3.788642786,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.458944 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a6cd6e5d1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.355150801 +0000 UTC m=+3.803846197,LastTimestamp:2026-03-12 13:09:45.355150801 +0000 UTC m=+3.803846197,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.465577 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a7815cad1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.543822033 +0000 UTC m=+3.992517429,LastTimestamp:2026-03-12 13:09:45.543822033 +0000 UTC m=+3.992517429,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.473482 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0a79948cc9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.568906441 +0000 UTC m=+4.017601837,LastTimestamp:2026-03-12 13:09:45.568906441 +0000 UTC m=+4.017601837,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.482447 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0aa714b647 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.332280391 +0000 UTC m=+4.780975827,LastTimestamp:2026-03-12 13:09:46.332280391 +0000 UTC m=+4.780975827,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.489040 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0ab151033a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.50400441 +0000 UTC m=+4.952699806,LastTimestamp:2026-03-12 13:09:46.50400441 +0000 UTC m=+4.952699806,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.494682 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0ab21c20fa openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.517315834 +0000 UTC m=+4.966011230,LastTimestamp:2026-03-12 13:09:46.517315834 +0000 UTC m=+4.966011230,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.501197 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0ab22eeac2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.518547138 +0000 UTC m=+4.967242534,LastTimestamp:2026-03-12 13:09:46.518547138 +0000 UTC m=+4.967242534,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.507315 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0abceacfcb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.698633163 +0000 UTC m=+5.147328559,LastTimestamp:2026-03-12 13:09:46.698633163 +0000 UTC m=+5.147328559,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.512382 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0abe0ad0fc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.717507836 +0000 UTC m=+5.166203222,LastTimestamp:2026-03-12 13:09:46.717507836 +0000 UTC m=+5.166203222,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.518830 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0abe1fdad0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.718886608 +0000 UTC m=+5.167582004,LastTimestamp:2026-03-12 13:09:46.718886608 +0000 UTC m=+5.167582004,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.523030 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0acbaf2c42 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.946382914 +0000 UTC m=+5.395078350,LastTimestamp:2026-03-12 13:09:46.946382914 +0000 UTC m=+5.395078350,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.526943 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0acc9757db openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.961598427 +0000 UTC m=+5.410293843,LastTimestamp:2026-03-12 13:09:46.961598427 +0000 UTC m=+5.410293843,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.531214 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0accb38f8d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:46.963447693 +0000 UTC m=+5.412143129,LastTimestamp:2026-03-12 13:09:46.963447693 +0000 UTC m=+5.412143129,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.535233 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0ada258c5c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:47.189021788 +0000 UTC m=+5.637717194,LastTimestamp:2026-03-12 13:09:47.189021788 +0000 UTC m=+5.637717194,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.541091 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0adb37edfd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:47.207003645 +0000 UTC m=+5.655699081,LastTimestamp:2026-03-12 13:09:47.207003645 +0000 UTC m=+5.655699081,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.545536 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0adb5277c8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:47.208742856 +0000 UTC m=+5.657438302,LastTimestamp:2026-03-12 13:09:47.208742856 +0000 UTC m=+5.657438302,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.549356 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0ae761d5da openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:47.41107657 +0000 UTC m=+5.859771966,LastTimestamp:2026-03-12 13:09:47.41107657 +0000 UTC m=+5.859771966,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.552470 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c1a0ae80be07a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:47.42222041 +0000 UTC m=+5.870915806,LastTimestamp:2026-03-12 13:09:47.42222041 +0000 UTC m=+5.870915806,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.556943 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 13:10:26 crc kubenswrapper[4778]: &Event{ObjectMeta:{kube-controller-manager-crc.189c1a0cb6fcc016 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 12 13:10:26 crc kubenswrapper[4778]: body: Mar 12 13:10:26 crc kubenswrapper[4778]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:55.189080086 +0000 UTC m=+13.637775492,LastTimestamp:2026-03-12 13:09:55.189080086 +0000 UTC m=+13.637775492,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 13:10:26 crc kubenswrapper[4778]: > Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.561144 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0cb6fdddec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:55.18915326 +0000 UTC m=+13.637848676,LastTimestamp:2026-03-12 13:09:55.18915326 +0000 UTC m=+13.637848676,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.564329 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 12 13:10:26 crc kubenswrapper[4778]: &Event{ObjectMeta:{kube-apiserver-crc.189c1a0ce8134087 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 12 13:10:26 crc kubenswrapper[4778]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 13:10:26 crc kubenswrapper[4778]: Mar 12 13:10:26 crc kubenswrapper[4778]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:56.012638343 +0000 UTC m=+14.461333749,LastTimestamp:2026-03-12 13:09:56.012638343 +0000 UTC m=+14.461333749,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 13:10:26 crc kubenswrapper[4778]: > Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.567314 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0ce814c327 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:56.012737319 +0000 UTC m=+14.461432725,LastTimestamp:2026-03-12 13:09:56.012737319 +0000 UTC m=+14.461432725,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.570469 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c1a0ce8134087\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 12 13:10:26 crc kubenswrapper[4778]: &Event{ObjectMeta:{kube-apiserver-crc.189c1a0ce8134087 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 12 13:10:26 crc kubenswrapper[4778]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 13:10:26 crc kubenswrapper[4778]: Mar 12 13:10:26 crc kubenswrapper[4778]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:56.012638343 +0000 UTC m=+14.461333749,LastTimestamp:2026-03-12 13:09:56.027862377 +0000 UTC m=+14.476557783,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 13:10:26 crc kubenswrapper[4778]: > Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.573658 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c1a0ce814c327\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0ce814c327 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:56.012737319 +0000 UTC m=+14.461432725,LastTimestamp:2026-03-12 13:09:56.027908029 +0000 UTC m=+14.476603435,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.577157 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c1a0a60a53c9f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a60a53c9f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.150569631 +0000 UTC m=+3.599265027,LastTimestamp:2026-03-12 13:09:56.362921586 +0000 UTC m=+14.811616982,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.581076 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c1a0a6beee97e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a6beee97e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.33994739 +0000 UTC m=+3.788642786,LastTimestamp:2026-03-12 13:09:56.537305644 +0000 UTC m=+14.986001040,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.584389 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c1a0a6cd6e5d1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c1a0a6cd6e5d1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:45.355150801 +0000 UTC m=+3.803846197,LastTimestamp:2026-03-12 13:09:56.543341638 +0000 UTC m=+14.992037034,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.588661 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 13:10:26 crc kubenswrapper[4778]: &Event{ObjectMeta:{kube-controller-manager-crc.189c1a0f0b10ac88 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 13:10:26 crc kubenswrapper[4778]: body: Mar 12 13:10:26 crc kubenswrapper[4778]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:05.189606536 +0000 UTC m=+23.638301972,LastTimestamp:2026-03-12 13:10:05.189606536 +0000 UTC m=+23.638301972,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 13:10:26 crc kubenswrapper[4778]: > Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.591805 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0f0b12176e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:05.189699438 +0000 UTC m=+23.638394874,LastTimestamp:2026-03-12 13:10:05.189699438 +0000 UTC m=+23.638394874,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.593214 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 13:10:26 crc kubenswrapper[4778]: &Event{ObjectMeta:{kube-controller-manager-crc.189c1a115ba16f46 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:47382->192.168.126.11:10357: read: connection reset by peer Mar 12 13:10:26 crc kubenswrapper[4778]: body: Mar 12 13:10:26 crc kubenswrapper[4778]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:15.131205446 +0000 UTC m=+33.579900882,LastTimestamp:2026-03-12 13:10:15.131205446 +0000 UTC m=+33.579900882,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 13:10:26 crc kubenswrapper[4778]: > Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.595597 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a115ba26d12 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:47382->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:15.131270418 +0000 UTC m=+33.579965854,LastTimestamp:2026-03-12 13:10:15.131270418 +0000 UTC m=+33.579965854,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.598594 4778 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a115bca7578 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:15.133894008 +0000 UTC m=+33.582589444,LastTimestamp:2026-03-12 13:10:15.133894008 +0000 UTC m=+33.582589444,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.601918 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c1a0a0647c085\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a0647c085 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.634493573 +0000 UTC m=+2.083188979,LastTimestamp:2026-03-12 13:10:15.165684508 +0000 UTC m=+33.614379944,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.605102 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c1a0a196f74ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a196f74ff openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.955862783 +0000 UTC m=+2.404558219,LastTimestamp:2026-03-12 13:10:15.382984139 +0000 UTC m=+33.831679545,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.608323 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c1a0a1af68cfa\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0a1af68cfa openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:09:43.981493498 +0000 UTC m=+2.430188924,LastTimestamp:2026-03-12 13:10:15.392822862 +0000 UTC m=+33.841518258,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.612817 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c1a0f0b10ac88\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 13:10:26 crc kubenswrapper[4778]: &Event{ObjectMeta:{kube-controller-manager-crc.189c1a0f0b10ac88 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 13:10:26 crc kubenswrapper[4778]: body: Mar 12 13:10:26 crc kubenswrapper[4778]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:05.189606536 +0000 UTC m=+23.638301972,LastTimestamp:2026-03-12 13:10:25.18997446 +0000 UTC m=+43.638669886,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 13:10:26 crc kubenswrapper[4778]: > Mar 12 13:10:26 crc kubenswrapper[4778]: E0312 13:10:26.616861 4778 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c1a0f0b12176e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c1a0f0b12176e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:10:05.189699438 +0000 UTC m=+23.638394874,LastTimestamp:2026-03-12 13:10:25.190036202 +0000 UTC m=+43.638731628,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:10:27 crc kubenswrapper[4778]: I0312 13:10:27.203937 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.029302 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.029536 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.030707 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.030746 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.030761 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.031316 4778 scope.go:117] "RemoveContainer" containerID="d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d" Mar 12 13:10:28 crc kubenswrapper[4778]: E0312 13:10:28.031492 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:10:28 crc kubenswrapper[4778]: I0312 13:10:28.202236 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:28 crc kubenswrapper[4778]: W0312 13:10:28.569011 4778 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:28 crc kubenswrapper[4778]: E0312 13:10:28.569094 4778 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 13:10:29 crc kubenswrapper[4778]: I0312 13:10:29.203037 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:29 crc kubenswrapper[4778]: W0312 13:10:29.343416 4778 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 12 13:10:29 crc kubenswrapper[4778]: E0312 13:10:29.343478 4778 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 12 13:10:30 crc kubenswrapper[4778]: I0312 13:10:30.202064 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:30 crc kubenswrapper[4778]: I0312 13:10:30.426703 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:30 crc kubenswrapper[4778]: I0312 13:10:30.428291 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:30 crc kubenswrapper[4778]: I0312 13:10:30.428919 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:30 crc kubenswrapper[4778]: I0312 13:10:30.428947 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:30 crc kubenswrapper[4778]: I0312 13:10:30.428983 4778 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 13:10:30 crc kubenswrapper[4778]: E0312 13:10:30.434971 4778 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 13:10:30 crc kubenswrapper[4778]: E0312 13:10:30.443245 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 13:10:31 crc kubenswrapper[4778]: I0312 13:10:31.203607 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.193709 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.194066 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.195143 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.195176 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.195207 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.197440 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.200920 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:32 crc kubenswrapper[4778]: E0312 13:10:32.357143 4778 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.484251 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.485005 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.485068 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:32 crc kubenswrapper[4778]: I0312 13:10:32.485079 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:33 crc kubenswrapper[4778]: I0312 13:10:33.201594 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:34 crc kubenswrapper[4778]: I0312 13:10:34.203417 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:35 crc kubenswrapper[4778]: I0312 13:10:35.202616 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:36 crc kubenswrapper[4778]: I0312 13:10:36.204492 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:36 crc kubenswrapper[4778]: W0312 13:10:36.832326 4778 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 12 13:10:36 crc kubenswrapper[4778]: E0312 13:10:36.832679 4778 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 13:10:36 crc kubenswrapper[4778]: I0312 13:10:36.836170 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 13:10:36 crc kubenswrapper[4778]: I0312 13:10:36.836418 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:36 crc kubenswrapper[4778]: I0312 13:10:36.837742 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:36 crc kubenswrapper[4778]: I0312 13:10:36.837780 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:36 crc kubenswrapper[4778]: I0312 13:10:36.837789 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:37 crc kubenswrapper[4778]: I0312 13:10:37.204828 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:37 crc kubenswrapper[4778]: I0312 13:10:37.435945 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:37 crc kubenswrapper[4778]: I0312 13:10:37.437573 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:37 crc kubenswrapper[4778]: I0312 13:10:37.437680 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:37 crc kubenswrapper[4778]: I0312 13:10:37.437699 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:37 crc kubenswrapper[4778]: I0312 13:10:37.437733 4778 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 13:10:37 crc kubenswrapper[4778]: E0312 13:10:37.443037 4778 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 13:10:37 crc kubenswrapper[4778]: E0312 13:10:37.450808 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 13:10:38 crc kubenswrapper[4778]: I0312 13:10:38.204066 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:39 crc kubenswrapper[4778]: I0312 13:10:39.202225 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:40 crc kubenswrapper[4778]: I0312 13:10:40.203417 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:41 crc kubenswrapper[4778]: I0312 13:10:41.201373 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.203518 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.252940 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.254136 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.254454 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.254507 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.255136 4778 scope.go:117] "RemoveContainer" containerID="d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d" Mar 12 13:10:42 crc kubenswrapper[4778]: E0312 13:10:42.357331 4778 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.512251 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.514222 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd"} Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.514419 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.515609 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.515644 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:42 crc kubenswrapper[4778]: I0312 13:10:42.515654 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.201825 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.518678 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.519137 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.521480 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" exitCode=255 Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.521558 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd"} Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.521609 4778 scope.go:117] "RemoveContainer" containerID="d1bfe956cf856eb5d8ec2a24a00da9f1dfcad215016a56650f0b3dec0dffaa4d" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.521837 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.523982 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.524012 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.524021 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.524493 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:10:43 crc kubenswrapper[4778]: E0312 13:10:43.524640 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:10:43 crc kubenswrapper[4778]: I0312 13:10:43.534947 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.203455 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.444117 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.445468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.445503 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.445512 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.445534 4778 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 13:10:44 crc kubenswrapper[4778]: E0312 13:10:44.450040 4778 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 13:10:44 crc kubenswrapper[4778]: E0312 13:10:44.455194 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.525347 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.526943 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.527667 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.527708 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.527722 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:44 crc kubenswrapper[4778]: I0312 13:10:44.528328 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:10:44 crc kubenswrapper[4778]: E0312 13:10:44.528524 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:10:45 crc kubenswrapper[4778]: I0312 13:10:45.205405 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:46 crc kubenswrapper[4778]: I0312 13:10:46.200928 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:47 crc kubenswrapper[4778]: I0312 13:10:47.201455 4778 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 13:10:47 crc kubenswrapper[4778]: I0312 13:10:47.207452 4778 csr.go:261] certificate signing request csr-zkwnc is approved, waiting to be issued Mar 12 13:10:47 crc kubenswrapper[4778]: I0312 13:10:47.217109 4778 csr.go:257] certificate signing request csr-zkwnc is issued Mar 12 13:10:47 crc kubenswrapper[4778]: I0312 13:10:47.235891 4778 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.029709 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.029879 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.030903 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.030936 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.030945 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.031493 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:10:48 crc kubenswrapper[4778]: E0312 13:10:48.031663 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.048723 4778 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.218114 4778 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-26 20:54:44.158263742 +0000 UTC Mar 12 13:10:48 crc kubenswrapper[4778]: I0312 13:10:48.218152 4778 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6223h43m55.940115094s for next certificate rotation Mar 12 13:10:49 crc kubenswrapper[4778]: I0312 13:10:49.877473 4778 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.450913 4778 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.452409 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.452450 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.452463 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.452587 4778 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.465079 4778 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.465389 4778 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.465418 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.469785 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.469831 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.469853 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.469884 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.469906 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:51Z","lastTransitionTime":"2026-03-12T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.488360 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.492436 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.492472 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.492481 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.492495 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.492511 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:51Z","lastTransitionTime":"2026-03-12T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.502689 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.505652 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.505679 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.505691 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.505706 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.505719 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:51Z","lastTransitionTime":"2026-03-12T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.519346 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.523797 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.523834 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.523843 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.523858 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:51 crc kubenswrapper[4778]: I0312 13:10:51.523868 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:51Z","lastTransitionTime":"2026-03-12T13:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.536886 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.537291 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.537386 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.637900 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.738800 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.839600 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:51 crc kubenswrapper[4778]: E0312 13:10:51.940601 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.040944 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.141463 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.241942 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.342647 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.357431 4778 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.443155 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.543821 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.644237 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.744613 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.845622 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:52 crc kubenswrapper[4778]: E0312 13:10:52.946334 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.047090 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.147570 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.247839 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.349002 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.450174 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.551127 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.652160 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.752492 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.853353 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:53 crc kubenswrapper[4778]: E0312 13:10:53.953747 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.053873 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.154258 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.255443 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.355810 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.456154 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.556705 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.657602 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.758392 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.859246 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:54 crc kubenswrapper[4778]: E0312 13:10:54.959935 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.060839 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.161895 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.262837 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.363565 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.464003 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.564596 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.665090 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.765526 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.866698 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:55 crc kubenswrapper[4778]: E0312 13:10:55.967300 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.067743 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.168629 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.269245 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.370359 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.471264 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.572084 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.672888 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.773819 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.874646 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:56 crc kubenswrapper[4778]: E0312 13:10:56.975416 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: E0312 13:10:57.075948 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: E0312 13:10:57.176279 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: E0312 13:10:57.276714 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: E0312 13:10:57.377361 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: E0312 13:10:57.478436 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: E0312 13:10:57.579558 4778 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.647897 4778 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.682634 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.682719 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.682745 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.682780 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.682805 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:57Z","lastTransitionTime":"2026-03-12T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.785372 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.785430 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.785451 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.785478 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.785504 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:57Z","lastTransitionTime":"2026-03-12T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.888071 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.888117 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.888127 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.888140 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.888152 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:57Z","lastTransitionTime":"2026-03-12T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.990301 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.990355 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.990370 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.990385 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:57 crc kubenswrapper[4778]: I0312 13:10:57.990397 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:57Z","lastTransitionTime":"2026-03-12T13:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.092214 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.092252 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.092260 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.092275 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.092284 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.194300 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.194334 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.194346 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.194362 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.194374 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.218079 4778 apiserver.go:52] "Watching apiserver" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.226530 4778 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227035 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qdxm2","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j","openshift-image-registry/node-ca-4dfhs","openshift-machine-config-operator/machine-config-daemon-2qx88","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-node-8bcc9","openshift-multus/multus-fhcz6","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-multus/multus-additional-cni-plugins-rsshp","openshift-multus/network-metrics-daemon-rz9vw","openshift-network-diagnostics/network-check-target-xd92c"] Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227637 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227755 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227810 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227846 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227883 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.227956 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.228133 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.228142 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.228432 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.228457 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.229247 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.229276 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.229246 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.229337 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.229445 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.230877 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.231068 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.231227 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.231363 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.232148 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.232767 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.232892 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.233078 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.235467 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.238552 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.241067 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.241723 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.241805 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.241812 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.241807 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.243965 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244107 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244131 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244155 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244194 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244266 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244271 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244323 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244356 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244378 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244384 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244386 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244496 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244505 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244599 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244624 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244665 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244685 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244723 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244507 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244815 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244846 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244862 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.244941 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.245303 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.261320 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.276618 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.284413 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.295284 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.295924 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.295970 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.295980 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.295996 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.296005 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.302594 4778 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.305276 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.308555 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.308707 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.308812 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.309391 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.309518 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310164 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.309093 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.309318 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.309456 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.309701 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310234 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310355 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310399 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310428 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310457 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310504 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310534 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310564 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310593 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310618 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310644 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310670 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310688 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310700 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310729 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310758 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310787 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.310910 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311146 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311363 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311514 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311562 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311568 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311584 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311590 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311636 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311657 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311676 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311693 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311709 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311727 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311748 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311765 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311782 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311800 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311817 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311833 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311849 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311867 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311921 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311942 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311960 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311977 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312008 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312024 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312044 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312062 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312080 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312096 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312115 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312131 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312148 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312163 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312180 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312210 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312228 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312244 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312259 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312279 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312295 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312331 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312350 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312366 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312383 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312400 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312415 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312432 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312447 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312462 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312477 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312493 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312510 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312524 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312584 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312603 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312619 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312635 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312655 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312669 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312684 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312701 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312716 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312732 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312749 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312765 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312783 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312799 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312813 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312829 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312846 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312863 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312880 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312896 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312914 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312931 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312947 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312964 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312979 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312996 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313012 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313028 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313043 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313060 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313076 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313092 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313109 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313167 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313195 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313214 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313241 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313260 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313276 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313294 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313311 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313330 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313347 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313363 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313379 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313394 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313411 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313427 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313443 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313460 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313476 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313493 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313510 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313526 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313541 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313559 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313574 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313591 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313609 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313624 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313642 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313659 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313675 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313693 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313709 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313729 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313745 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313762 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313778 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313795 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313812 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313829 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313846 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313862 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313879 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313899 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313916 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313932 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313949 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313966 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313981 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313999 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314015 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314034 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314050 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314067 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314084 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314101 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314117 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314134 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314151 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314168 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314218 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314240 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314257 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314273 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314290 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314310 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314326 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314343 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314360 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314378 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314394 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314412 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314429 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314446 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314462 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314478 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314498 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314519 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314543 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314567 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314588 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314609 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314626 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314646 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314664 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314682 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314706 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314732 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314752 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314770 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314787 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.311720 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314854 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314892 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-cni-binary-copy\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314909 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-k8s-cni-cncf-io\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314926 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314945 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314962 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314980 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-os-release\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315001 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315018 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312052 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312250 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.312744 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313615 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313631 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313661 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313390 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.313930 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314342 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314449 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.314841 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315039 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4hrx\" (UniqueName: \"kubernetes.io/projected/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-kube-api-access-w4hrx\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315385 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-conf-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315428 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-ovn-kubernetes\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315460 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-netd\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315499 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/24438fc6-dab0-4a9e-8b97-2532da76d9cd-rootfs\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315527 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfefcab6-a931-413e-8763-0f63f17911cd-host\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.315541 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:10:58.815521159 +0000 UTC m=+77.264216555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315571 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g92p9\" (UniqueName: \"kubernetes.io/projected/de004a2f-3061-4aae-aa57-389219c71023-kube-api-access-g92p9\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315598 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdj5w\" (UniqueName: \"kubernetes.io/projected/0b59b25a-3acc-4d06-b91d-575f45463520-kube-api-access-cdj5w\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315618 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-config\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315652 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-slash\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315670 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cnibin\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315689 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315707 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-cni-bin\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315725 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-netns\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315742 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/de004a2f-3061-4aae-aa57-389219c71023-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315759 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315786 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/24438fc6-dab0-4a9e-8b97-2532da76d9cd-mcd-auth-proxy-config\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315809 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/de004a2f-3061-4aae-aa57-389219c71023-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315829 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-762lp\" (UniqueName: \"kubernetes.io/projected/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-kube-api-access-762lp\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315848 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-var-lib-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315866 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/de004a2f-3061-4aae-aa57-389219c71023-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315894 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315911 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-os-release\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315932 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315948 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-systemd\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315964 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-node-log\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.315979 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-log-socket\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316000 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316017 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhn9v\" (UniqueName: \"kubernetes.io/projected/24438fc6-dab0-4a9e-8b97-2532da76d9cd-kube-api-access-rhn9v\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316035 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-system-cni-dir\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316053 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-etc-kubernetes\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316069 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-kubelet\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316085 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-bin\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316102 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/24438fc6-dab0-4a9e-8b97-2532da76d9cd-proxy-tls\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316118 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cfefcab6-a931-413e-8763-0f63f17911cd-serviceca\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316133 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316148 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-cni-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316167 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-env-overrides\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316204 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-script-lib\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316234 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316258 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7050ddd9-aa01-4af7-9046-208f85f50a86-hosts-file\") pod \"node-resolver-qdxm2\" (UID: \"7050ddd9-aa01-4af7-9046-208f85f50a86\") " pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316279 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-cnibin\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316296 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-socket-dir-parent\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316311 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-netns\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316331 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-cni-multus\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316353 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-daemon-config\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316377 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316401 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316430 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-ovn\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316454 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-schvw\" (UniqueName: \"kubernetes.io/projected/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-kube-api-access-schvw\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316491 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316567 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316518 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssbrl\" (UniqueName: \"kubernetes.io/projected/cfefcab6-a931-413e-8763-0f63f17911cd-kube-api-access-ssbrl\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316710 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jspwn\" (UniqueName: \"kubernetes.io/projected/7050ddd9-aa01-4af7-9046-208f85f50a86-kube-api-access-jspwn\") pod \"node-resolver-qdxm2\" (UID: \"7050ddd9-aa01-4af7-9046-208f85f50a86\") " pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316736 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-system-cni-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316738 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-kubelet\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316787 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-hostroot\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316809 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-systemd-units\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316834 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-etc-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316838 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316857 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovn-node-metrics-cert\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316887 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316888 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.316924 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.317404 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.317898 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.317976 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318017 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318324 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318483 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318549 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318691 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318775 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318843 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318856 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318897 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.318915 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319084 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319167 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319293 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319721 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319740 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319756 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320004 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320029 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320028 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.319788 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320139 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320176 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320231 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320277 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.320833 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.321346 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.321400 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.321550 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.322006 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.322783 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-multus-certs\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.322875 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.322971 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.323167 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.323341 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:58.823274799 +0000 UTC m=+77.271970235 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323360 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323534 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323539 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323723 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323888 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323910 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.324628 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.325945 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.326062 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.323540 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.326274 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.326346 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.326460 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.326592 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:58.826574015 +0000 UTC m=+77.275269461 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.326901 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328344 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328401 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328431 4778 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328610 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328651 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328669 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329017 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328692 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328950 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329062 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.328961 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329091 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329673 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329707 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329828 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.329873 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.330020 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.330363 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.330825 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.330960 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331089 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331250 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331509 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331735 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331860 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331811 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331924 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.331929 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332261 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332437 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332513 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332467 4778 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332670 4778 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332673 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332701 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332707 4778 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332931 4778 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332952 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332963 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332974 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332985 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.332996 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333053 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333087 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333101 4778 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333112 4778 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333123 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333137 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333146 4778 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333175 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333206 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333217 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333229 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333232 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333242 4778 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333256 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333290 4778 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333302 4778 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333311 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333323 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333332 4778 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333557 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.333760 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.334243 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.341694 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.343112 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.343154 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.343169 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.343261 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:58.843239015 +0000 UTC m=+77.291934411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.343550 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.344201 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.346434 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.347491 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.347604 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.347679 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.347738 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.347762 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.347717 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.348745 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:58.848705376 +0000 UTC m=+77.297400792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.348832 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.354738 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.354797 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.355234 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.355252 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.355800 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356200 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356238 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356350 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356673 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356833 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356927 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.357025 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.357482 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.356932 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.358764 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.359567 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.359849 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.359983 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360211 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360217 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362525 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360564 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360616 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360715 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360888 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.360888 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.361140 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.361216 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.361285 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362212 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362642 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.361911 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.361928 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362953 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362253 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362373 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362487 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362496 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362534 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.362652 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.364465 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.366523 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.366657 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.366900 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.366910 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.367043 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.367220 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.367504 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.367521 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.368047 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.368289 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.369525 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.369879 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.369947 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.370107 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.370668 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371000 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371318 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371374 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371541 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371626 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371687 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.371917 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372015 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372093 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372109 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372411 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372663 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372816 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.372859 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373269 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373297 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373323 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373334 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373375 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373382 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373522 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373549 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373927 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.373974 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.374075 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.374329 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.374429 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.374781 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.380404 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.380699 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.386227 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.391590 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.394925 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.399691 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.399754 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.399768 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.399786 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.399800 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.404650 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.406781 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.408398 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.413248 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.421949 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.432966 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434227 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/24438fc6-dab0-4a9e-8b97-2532da76d9cd-mcd-auth-proxy-config\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434257 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/de004a2f-3061-4aae-aa57-389219c71023-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434275 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-762lp\" (UniqueName: \"kubernetes.io/projected/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-kube-api-access-762lp\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434292 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-var-lib-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434307 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-systemd\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434323 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-node-log\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434339 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/de004a2f-3061-4aae-aa57-389219c71023-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434362 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-os-release\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434377 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-log-socket\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434391 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-kubelet\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434404 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-bin\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434425 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhn9v\" (UniqueName: \"kubernetes.io/projected/24438fc6-dab0-4a9e-8b97-2532da76d9cd-kube-api-access-rhn9v\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434440 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-system-cni-dir\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434455 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-etc-kubernetes\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434469 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-script-lib\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434483 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/24438fc6-dab0-4a9e-8b97-2532da76d9cd-proxy-tls\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434497 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cfefcab6-a931-413e-8763-0f63f17911cd-serviceca\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434513 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434527 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-cni-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434541 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-env-overrides\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434555 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7050ddd9-aa01-4af7-9046-208f85f50a86-hosts-file\") pod \"node-resolver-qdxm2\" (UID: \"7050ddd9-aa01-4af7-9046-208f85f50a86\") " pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434568 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-cnibin\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434581 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-socket-dir-parent\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434597 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434610 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-ovn\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434624 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-schvw\" (UniqueName: \"kubernetes.io/projected/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-kube-api-access-schvw\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434639 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-netns\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434653 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-cni-multus\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434667 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-daemon-config\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434681 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-hostroot\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434695 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-systemd-units\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435294 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-etc-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434740 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-etc-kubernetes\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.434788 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.435502 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:58.935482845 +0000 UTC m=+77.384178241 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434867 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-systemd-units\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434909 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-socket-dir-parent\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435526 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-env-overrides\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434950 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-ovn\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435560 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-etc-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435169 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-netns\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435222 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-cni-multus\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434931 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435580 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-script-lib\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435564 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/24438fc6-dab0-4a9e-8b97-2532da76d9cd-mcd-auth-proxy-config\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434777 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-cnibin\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435626 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-os-release\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.434839 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-cni-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435674 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7050ddd9-aa01-4af7-9046-208f85f50a86-hosts-file\") pod \"node-resolver-qdxm2\" (UID: \"7050ddd9-aa01-4af7-9046-208f85f50a86\") " pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435694 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssbrl\" (UniqueName: \"kubernetes.io/projected/cfefcab6-a931-413e-8763-0f63f17911cd-kube-api-access-ssbrl\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435720 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jspwn\" (UniqueName: \"kubernetes.io/projected/7050ddd9-aa01-4af7-9046-208f85f50a86-kube-api-access-jspwn\") pod \"node-resolver-qdxm2\" (UID: \"7050ddd9-aa01-4af7-9046-208f85f50a86\") " pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435741 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-system-cni-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435761 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-kubelet\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435781 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovn-node-metrics-cert\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435802 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435827 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435848 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-multus-certs\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435857 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-daemon-config\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435878 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-cni-binary-copy\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435898 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-k8s-cni-cncf-io\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435902 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-hostroot\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435918 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435947 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435967 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-os-release\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435682 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-log-socket\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.435989 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4hrx\" (UniqueName: \"kubernetes.io/projected/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-kube-api-access-w4hrx\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436016 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-conf-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436041 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-ovn-kubernetes\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436066 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-netd\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436093 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdj5w\" (UniqueName: \"kubernetes.io/projected/0b59b25a-3acc-4d06-b91d-575f45463520-kube-api-access-cdj5w\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436115 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-config\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436137 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/24438fc6-dab0-4a9e-8b97-2532da76d9cd-rootfs\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436156 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfefcab6-a931-413e-8763-0f63f17911cd-host\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436177 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g92p9\" (UniqueName: \"kubernetes.io/projected/de004a2f-3061-4aae-aa57-389219c71023-kube-api-access-g92p9\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436224 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-slash\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436246 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cnibin\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436272 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-systemd\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436315 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-bin\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436317 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-node-log\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436358 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-system-cni-dir\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436359 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-kubelet\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436089 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-var-lib-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436411 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-os-release\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436451 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-ovn-kubernetes\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436506 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-netd\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436517 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-multus-conf-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436620 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfefcab6-a931-413e-8763-0f63f17911cd-host\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436656 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/24438fc6-dab0-4a9e-8b97-2532da76d9cd-rootfs\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436687 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-slash\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436688 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-multus-certs\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436735 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-kubelet\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436796 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-system-cni-dir\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436852 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436268 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436989 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/de004a2f-3061-4aae-aa57-389219c71023-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.436974 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-cni-bin\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437050 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-netns\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437052 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-run-k8s-cni-cncf-io\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437070 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/de004a2f-3061-4aae-aa57-389219c71023-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437096 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437168 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437202 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437218 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437233 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437248 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437260 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437272 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437273 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-openvswitch\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437285 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437317 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437322 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437337 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437348 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cnibin\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437351 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437372 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437385 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437396 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437408 4778 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437421 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437433 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437446 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437458 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437470 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437443 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-host-var-lib-cni-bin\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437500 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-netns\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437482 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437621 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437626 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437714 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-cni-binary-copy\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437636 4778 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437574 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-config\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437767 4778 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437790 4778 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437805 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437821 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.437836 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.438105 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cfefcab6-a931-413e-8763-0f63f17911cd-serviceca\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.438139 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/de004a2f-3061-4aae-aa57-389219c71023-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.439067 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440738 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440767 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440781 4778 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440800 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440812 4778 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440898 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440914 4778 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440927 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440944 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.440965 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.441238 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442489 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442510 4778 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442529 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442543 4778 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442558 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442571 4778 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442589 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442653 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442664 4778 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442677 4778 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442689 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442700 4778 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442709 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442719 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442729 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442738 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442747 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442756 4778 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442767 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442777 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442787 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442797 4778 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442806 4778 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442815 4778 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442824 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442833 4778 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442842 4778 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442851 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442860 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442868 4778 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442877 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442887 4778 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442895 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442904 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442912 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442922 4778 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442931 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442941 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442950 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442959 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442967 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442976 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442985 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.442993 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443002 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443011 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443020 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443030 4778 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443038 4778 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443048 4778 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443056 4778 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443064 4778 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443072 4778 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443082 4778 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443090 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443098 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443110 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443119 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443129 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443138 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443147 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443155 4778 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443163 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443172 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443201 4778 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443213 4778 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443222 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443230 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443240 4778 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443249 4778 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443257 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443266 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443275 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443283 4778 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443292 4778 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443327 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443337 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443345 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443354 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443363 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443373 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443381 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443390 4778 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443400 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443409 4778 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443418 4778 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443427 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443436 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443445 4778 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443453 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443463 4778 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443476 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443484 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443494 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443504 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443513 4778 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443523 4778 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443531 4778 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443545 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443553 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443561 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443570 4778 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443579 4778 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443587 4778 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443595 4778 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443605 4778 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443613 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443622 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443632 4778 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443639 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443650 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443658 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443667 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443675 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443684 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443692 4778 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443701 4778 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443709 4778 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443718 4778 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443726 4778 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443735 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443743 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443750 4778 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443761 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443769 4778 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443777 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443786 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443795 4778 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.443803 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.444171 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovn-node-metrics-cert\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.447580 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/24438fc6-dab0-4a9e-8b97-2532da76d9cd-proxy-tls\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.452578 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/de004a2f-3061-4aae-aa57-389219c71023-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.454591 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssbrl\" (UniqueName: \"kubernetes.io/projected/cfefcab6-a931-413e-8763-0f63f17911cd-kube-api-access-ssbrl\") pod \"node-ca-4dfhs\" (UID: \"cfefcab6-a931-413e-8763-0f63f17911cd\") " pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.455248 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-schvw\" (UniqueName: \"kubernetes.io/projected/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-kube-api-access-schvw\") pod \"ovnkube-node-8bcc9\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.455990 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4hrx\" (UniqueName: \"kubernetes.io/projected/8a1f8eaa-ac07-4478-be5d-0742de6b43c4-kube-api-access-w4hrx\") pod \"multus-additional-cni-plugins-rsshp\" (UID: \"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\") " pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.456164 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdj5w\" (UniqueName: \"kubernetes.io/projected/0b59b25a-3acc-4d06-b91d-575f45463520-kube-api-access-cdj5w\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.456415 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-762lp\" (UniqueName: \"kubernetes.io/projected/1e7037a8-a966-4df0-9f94-fe2dd3e2de6e-kube-api-access-762lp\") pod \"multus-fhcz6\" (UID: \"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\") " pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.456942 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhn9v\" (UniqueName: \"kubernetes.io/projected/24438fc6-dab0-4a9e-8b97-2532da76d9cd-kube-api-access-rhn9v\") pod \"machine-config-daemon-2qx88\" (UID: \"24438fc6-dab0-4a9e-8b97-2532da76d9cd\") " pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.458126 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jspwn\" (UniqueName: \"kubernetes.io/projected/7050ddd9-aa01-4af7-9046-208f85f50a86-kube-api-access-jspwn\") pod \"node-resolver-qdxm2\" (UID: \"7050ddd9-aa01-4af7-9046-208f85f50a86\") " pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.459218 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g92p9\" (UniqueName: \"kubernetes.io/projected/de004a2f-3061-4aae-aa57-389219c71023-kube-api-access-g92p9\") pod \"ovnkube-control-plane-749d76644c-sww7j\" (UID: \"de004a2f-3061-4aae-aa57-389219c71023\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.502157 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.502224 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.502240 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.502261 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.502275 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.550325 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.556568 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.567411 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 13:10:58 crc kubenswrapper[4778]: W0312 13:10:58.569014 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-1eda1cd4295e4701797cceb7174e91b8d6499a94d1e66ba504e25839618a7acf WatchSource:0}: Error finding container 1eda1cd4295e4701797cceb7174e91b8d6499a94d1e66ba504e25839618a7acf: Status 404 returned error can't find the container with id 1eda1cd4295e4701797cceb7174e91b8d6499a94d1e66ba504e25839618a7acf Mar 12 13:10:58 crc kubenswrapper[4778]: W0312 13:10:58.570170 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24438fc6_dab0_4a9e_8b97_2532da76d9cd.slice/crio-9fb8cc13c65fc315644001720a3ed3b49ecd4b57157eaf28de06428d98a7432e WatchSource:0}: Error finding container 9fb8cc13c65fc315644001720a3ed3b49ecd4b57157eaf28de06428d98a7432e: Status 404 returned error can't find the container with id 9fb8cc13c65fc315644001720a3ed3b49ecd4b57157eaf28de06428d98a7432e Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.575618 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.584950 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:10:58 crc kubenswrapper[4778]: W0312 13:10:58.587472 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-1c02416320777aa48a416466c459c00560f0e4d96499a38198a553fa887aee3a WatchSource:0}: Error finding container 1c02416320777aa48a416466c459c00560f0e4d96499a38198a553fa887aee3a: Status 404 returned error can't find the container with id 1c02416320777aa48a416466c459c00560f0e4d96499a38198a553fa887aee3a Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.593360 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fhcz6" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.599419 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.604853 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qdxm2" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.605935 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.605972 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.605994 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.606020 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.606041 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: W0312 13:10:58.605926 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-4e8f021e062407652b69c8ca2911a1a25e12ea0074598fd72d17a5661a7e9d4d WatchSource:0}: Error finding container 4e8f021e062407652b69c8ca2911a1a25e12ea0074598fd72d17a5661a7e9d4d: Status 404 returned error can't find the container with id 4e8f021e062407652b69c8ca2911a1a25e12ea0074598fd72d17a5661a7e9d4d Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.614859 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4dfhs" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.617760 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rsshp" Mar 12 13:10:58 crc kubenswrapper[4778]: W0312 13:10:58.631291 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65cd795e_eb6e_4995_a4c1_9dea6f425ac5.slice/crio-591e87d9e47004fc9c6fc7b24484cec488177d8e0820b4787eb9618d9e5051df WatchSource:0}: Error finding container 591e87d9e47004fc9c6fc7b24484cec488177d8e0820b4787eb9618d9e5051df: Status 404 returned error can't find the container with id 591e87d9e47004fc9c6fc7b24484cec488177d8e0820b4787eb9618d9e5051df Mar 12 13:10:58 crc kubenswrapper[4778]: W0312 13:10:58.662220 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e7037a8_a966_4df0_9f94_fe2dd3e2de6e.slice/crio-99cbed59a9719cac7008be58a65d409090b8ab2da26045a2ec67cfe3d360061a WatchSource:0}: Error finding container 99cbed59a9719cac7008be58a65d409090b8ab2da26045a2ec67cfe3d360061a: Status 404 returned error can't find the container with id 99cbed59a9719cac7008be58a65d409090b8ab2da26045a2ec67cfe3d360061a Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.710774 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.710818 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.710832 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.710849 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.710863 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.813801 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.813831 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.813840 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.813854 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.813863 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.849141 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.849279 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849322 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:10:59.849289384 +0000 UTC m=+78.297984790 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849360 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.849368 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849419 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:59.849403667 +0000 UTC m=+78.298099123 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.849461 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.849493 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849562 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849579 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849591 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849596 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849600 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849652 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849663 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849630 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:59.849619353 +0000 UTC m=+78.298314989 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849708 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:59.849695985 +0000 UTC m=+78.298391381 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.849721 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:59.849713895 +0000 UTC m=+78.298409291 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.917261 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.917310 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.917318 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.917333 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.917913 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:58Z","lastTransitionTime":"2026-03-12T13:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.950207 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:10:58 crc kubenswrapper[4778]: I0312 13:10:58.950963 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:58 crc kubenswrapper[4778]: E0312 13:10:58.951030 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:10:59.951012749 +0000 UTC m=+78.399708145 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.022720 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.022874 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.022954 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.023041 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.023112 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.125805 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.125842 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.125850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.125864 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.125875 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.239607 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.239639 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.239650 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.239666 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.239681 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.342349 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.342414 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.342433 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.342454 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.342465 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.445028 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.445061 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.445071 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.445083 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.445398 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.548394 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.548468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.548486 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.548882 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.549120 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.566113 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1c02416320777aa48a416466c459c00560f0e4d96499a38198a553fa887aee3a"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.568229 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qdxm2" event={"ID":"7050ddd9-aa01-4af7-9046-208f85f50a86","Type":"ContainerStarted","Data":"9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.568266 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qdxm2" event={"ID":"7050ddd9-aa01-4af7-9046-208f85f50a86","Type":"ContainerStarted","Data":"ebdcb6fb90e54db80d802a165c170454f2513f546e2fcead9989de5a3a3734f4"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.570670 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fhcz6" event={"ID":"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e","Type":"ContainerStarted","Data":"5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.570727 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fhcz6" event={"ID":"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e","Type":"ContainerStarted","Data":"99cbed59a9719cac7008be58a65d409090b8ab2da26045a2ec67cfe3d360061a"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.572689 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.572737 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4e8f021e062407652b69c8ca2911a1a25e12ea0074598fd72d17a5661a7e9d4d"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.574484 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" event={"ID":"de004a2f-3061-4aae-aa57-389219c71023","Type":"ContainerStarted","Data":"d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.574515 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" event={"ID":"de004a2f-3061-4aae-aa57-389219c71023","Type":"ContainerStarted","Data":"478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.574528 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" event={"ID":"de004a2f-3061-4aae-aa57-389219c71023","Type":"ContainerStarted","Data":"96a2eabb09f2f7cd28d7f361a5a9eae956e1a74ad6b85c92f5ec73e7a303d94e"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.575919 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" exitCode=0 Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.576007 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.576040 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"591e87d9e47004fc9c6fc7b24484cec488177d8e0820b4787eb9618d9e5051df"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.578998 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.579047 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.579063 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1eda1cd4295e4701797cceb7174e91b8d6499a94d1e66ba504e25839618a7acf"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.582879 4778 generic.go:334] "Generic (PLEG): container finished" podID="8a1f8eaa-ac07-4478-be5d-0742de6b43c4" containerID="f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057" exitCode=0 Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.583022 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerDied","Data":"f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.583070 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerStarted","Data":"d9ed48aead91feac56fb2c82c53067915d8ab5047b52c7c84d4c0fa72f58c6ae"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.586824 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4dfhs" event={"ID":"cfefcab6-a931-413e-8763-0f63f17911cd","Type":"ContainerStarted","Data":"eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.587261 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4dfhs" event={"ID":"cfefcab6-a931-413e-8763-0f63f17911cd","Type":"ContainerStarted","Data":"7b6dfac531975a0fc0d0d91f8f34d5737bae2604221aa20505e2c947f0e29ecd"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.588298 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.588806 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.588892 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.588914 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"9fb8cc13c65fc315644001720a3ed3b49ecd4b57157eaf28de06428d98a7432e"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.606835 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.619515 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.633062 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.644371 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.653536 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.653561 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.653571 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.653585 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.653594 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.655160 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.671470 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.686525 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.698540 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.710204 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.720899 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.738305 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.747764 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.757938 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.757961 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.757987 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.758001 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.758013 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.761639 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.780356 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.790818 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.803717 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.815441 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.830507 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.844853 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.853730 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.860001 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.860036 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.860049 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.860065 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.860078 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.862511 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862680 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:11:01.862652486 +0000 UTC m=+80.311347882 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.862681 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.862762 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862776 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.862816 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862837 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:01.86282103 +0000 UTC m=+80.311516436 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.862860 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862881 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862917 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:01.862908463 +0000 UTC m=+80.311603859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862972 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862986 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862989 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.862999 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.863004 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.863017 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.863030 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:01.863020026 +0000 UTC m=+80.311715522 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.863045 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:01.863037066 +0000 UTC m=+80.311732552 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.865944 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.877242 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.887951 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.898243 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.910297 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.922004 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.935508 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:59Z is after 2025-08-24T17:21:41Z" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.962144 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.962220 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.962241 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.962264 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.962279 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:10:59Z","lastTransitionTime":"2026-03-12T13:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:10:59 crc kubenswrapper[4778]: I0312 13:10:59.963440 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.963538 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:10:59 crc kubenswrapper[4778]: E0312 13:10:59.963584 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:01.963572321 +0000 UTC m=+80.412267717 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.065128 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.065223 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.065248 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.065276 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.065297 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.168136 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.168204 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.168214 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.168228 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.168237 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.252955 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.253106 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.253102 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:00 crc kubenswrapper[4778]: E0312 13:11:00.253354 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.253510 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:00 crc kubenswrapper[4778]: E0312 13:11:00.253690 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:00 crc kubenswrapper[4778]: E0312 13:11:00.253840 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:00 crc kubenswrapper[4778]: E0312 13:11:00.254001 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.256936 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.257622 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.258848 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.259491 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.260491 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.261063 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.261681 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.262608 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.263283 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.264142 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.264654 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.265661 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.266137 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.266650 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.267565 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.268055 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.268958 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.269384 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.269936 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.271918 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.272174 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.272311 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.272334 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.272364 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.272388 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.273481 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.275073 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.276363 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.278305 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.279699 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.282176 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.283655 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.284475 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.285426 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.286207 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.286939 4778 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.287098 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.289079 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.289880 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.290638 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.293358 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.294536 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.296078 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.297039 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.298441 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.299491 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.300613 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.301740 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.302826 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.303685 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.305235 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.306046 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.307041 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.308210 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.310118 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.310856 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.312258 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.312918 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.314008 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.321730 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:11:00 crc kubenswrapper[4778]: E0312 13:11:00.321941 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.323416 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.374958 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.375017 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.375035 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.375061 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.375079 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.477731 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.477975 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.478124 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.478257 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.478352 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.580923 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.581388 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.581487 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.581616 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.581740 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.596278 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.596314 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.596324 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.596333 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.597930 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerStarted","Data":"deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.598807 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:11:00 crc kubenswrapper[4778]: E0312 13:11:00.599055 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.610773 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.625593 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.641439 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.658717 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.670995 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.683701 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.683742 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.683753 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.683769 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.683787 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.684887 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.695287 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.706563 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.727162 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.744894 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.772792 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.786174 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.786414 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.786481 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.786553 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.786630 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.787634 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.800709 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.825121 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.842856 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:00Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.890387 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.890427 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.890440 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.890459 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.890472 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.992717 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.992761 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.992771 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.992786 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:00 crc kubenswrapper[4778]: I0312 13:11:00.992797 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:00Z","lastTransitionTime":"2026-03-12T13:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.094702 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.094733 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.094749 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.094767 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.094778 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.196778 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.197049 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.197060 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.197097 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.197107 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.299755 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.300009 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.300104 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.300201 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.300270 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.403766 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.404003 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.404088 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.404163 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.404269 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.507522 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.507563 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.507577 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.507596 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.507608 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.603384 4778 generic.go:334] "Generic (PLEG): container finished" podID="8a1f8eaa-ac07-4478-be5d-0742de6b43c4" containerID="deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f" exitCode=0 Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.603505 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerDied","Data":"deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.608004 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.608042 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.608959 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.609000 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.609011 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.609030 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.609045 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.621378 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.635934 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.649089 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.660813 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.674112 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.688332 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.698492 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.711432 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.711485 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.711495 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.711509 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.711521 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.712509 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.731629 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.744077 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.757891 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.771260 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.778681 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.778711 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.778723 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.778739 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.778747 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.788680 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.789252 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.792795 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.792823 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.792834 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.792849 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.792860 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.801299 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.808824 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.810382 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.811946 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.811973 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.811984 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.812000 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.812011 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.822261 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.825929 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.825988 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.826003 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.826022 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.826035 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.836595 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.840663 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.840805 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.840888 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.840969 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.841030 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.855978 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:01Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.856103 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.857593 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.857633 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.857644 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.857656 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.857665 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.898171 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898340 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:11:05.89830846 +0000 UTC m=+84.347003866 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.898403 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.898521 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898580 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.898570 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898609 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898621 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.898652 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898667 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:05.898653479 +0000 UTC m=+84.347348875 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898731 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898767 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898813 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:05.898795302 +0000 UTC m=+84.347490698 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898836 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:05.898826993 +0000 UTC m=+84.347522479 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.898928 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.899006 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.899025 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.899121 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:05.89909594 +0000 UTC m=+84.347791336 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.959760 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.959814 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.959829 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.959850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.959864 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:01Z","lastTransitionTime":"2026-03-12T13:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:01 crc kubenswrapper[4778]: I0312 13:11:01.999340 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.999537 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:01 crc kubenswrapper[4778]: E0312 13:11:01.999619 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:05.999600544 +0000 UTC m=+84.448295940 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.063573 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.063609 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.063621 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.063639 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.063653 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.165705 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.165742 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.165753 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.165769 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.165780 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.253469 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.253489 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:02 crc kubenswrapper[4778]: E0312 13:11:02.253726 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.253761 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.253747 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:02 crc kubenswrapper[4778]: E0312 13:11:02.253898 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:02 crc kubenswrapper[4778]: E0312 13:11:02.254045 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:02 crc kubenswrapper[4778]: E0312 13:11:02.254250 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.268409 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.268479 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.268506 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.268538 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.268558 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.274468 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.290782 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.308654 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.327493 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.341908 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.367035 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.371052 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.371115 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.371133 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.371158 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.371175 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.393897 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.423608 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.435345 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.448324 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.458078 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.470079 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.473950 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.474138 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.474277 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.474391 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.474485 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.486637 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.501274 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.514980 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.577287 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.577334 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.577344 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.577365 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.577375 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.614998 4778 generic.go:334] "Generic (PLEG): container finished" podID="8a1f8eaa-ac07-4478-be5d-0742de6b43c4" containerID="2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028" exitCode=0 Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.615110 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerDied","Data":"2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.617158 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.634129 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.651873 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.668620 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.680534 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.680570 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.680581 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.680595 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.680606 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.685419 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.705390 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.718507 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.733084 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.749417 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.759974 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.774303 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.782666 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.782692 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.782700 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.782718 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.782729 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.790261 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.802086 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.813368 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.823804 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.838057 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.849900 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.860211 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.876644 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.886393 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.886433 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.886441 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.886455 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.886465 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.889955 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.901808 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.920014 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.939687 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.962275 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.988765 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.988805 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.988814 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.988826 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.988835 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:02Z","lastTransitionTime":"2026-03-12T13:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:02 crc kubenswrapper[4778]: I0312 13:11:02.990021 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:02.999963 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.015207 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.030019 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.047506 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.060364 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.072108 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.091308 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.091344 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.091352 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.091364 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.091373 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.193294 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.193347 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.193364 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.193387 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.193402 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.296086 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.296140 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.296157 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.296179 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.296237 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.398883 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.398936 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.398954 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.398977 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.398995 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.501362 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.501431 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.501455 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.501487 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.501508 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.603893 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.603935 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.603950 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.603970 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.603987 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.622395 4778 generic.go:334] "Generic (PLEG): container finished" podID="8a1f8eaa-ac07-4478-be5d-0742de6b43c4" containerID="3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d" exitCode=0 Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.622439 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerDied","Data":"3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.627456 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.641600 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.656260 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.671284 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.683425 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.697636 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.707472 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.707516 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.707533 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.707557 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.707573 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.709446 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.719354 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.736328 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.749432 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.760003 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.771419 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.783590 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.799589 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.809297 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.811691 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.811724 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.811735 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.811750 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.811763 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.826275 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.914722 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.914767 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.914781 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.914797 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:03 crc kubenswrapper[4778]: I0312 13:11:03.914809 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:03Z","lastTransitionTime":"2026-03-12T13:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.017871 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.017923 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.017946 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.017970 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.017987 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.122396 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.122468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.122489 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.122512 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.122533 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.225620 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.225697 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.225719 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.225749 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.225772 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.253650 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.253724 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.253893 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:04 crc kubenswrapper[4778]: E0312 13:11:04.253885 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.253936 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:04 crc kubenswrapper[4778]: E0312 13:11:04.254338 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:04 crc kubenswrapper[4778]: E0312 13:11:04.254527 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:04 crc kubenswrapper[4778]: E0312 13:11:04.254682 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.269269 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.329083 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.329132 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.329149 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.329174 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.329220 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.431727 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.431846 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.431867 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.431891 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.431909 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.535097 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.535179 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.535253 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.535297 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.535349 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.634220 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerStarted","Data":"253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.638468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.638529 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.638548 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.638574 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.638592 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.659915 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.686367 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.700440 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.717978 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.733276 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.740569 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.740620 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.740639 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.740660 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.740675 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.747595 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.760883 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.770745 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.783087 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.795923 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.810904 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.829274 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.842255 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.843374 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.843410 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.843424 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.843441 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.843453 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.856612 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.871802 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.884616 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:04Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.946240 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.946326 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.946344 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.946399 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:04 crc kubenswrapper[4778]: I0312 13:11:04.946417 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:04Z","lastTransitionTime":"2026-03-12T13:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.048779 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.048824 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.048834 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.048851 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.048860 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.151306 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.151355 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.151368 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.151385 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.151396 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.254024 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.254095 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.254123 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.254150 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.254175 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.357606 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.358034 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.358059 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.358081 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.358092 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.460618 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.460684 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.460703 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.460728 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.460748 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.563505 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.563545 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.563556 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.563574 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.563587 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.644324 4778 generic.go:334] "Generic (PLEG): container finished" podID="8a1f8eaa-ac07-4478-be5d-0742de6b43c4" containerID="253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1" exitCode=0 Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.644412 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerDied","Data":"253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.666366 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.666440 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.666462 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.666494 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.666518 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.667426 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.687766 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.703667 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.719947 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.735221 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.753597 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.766336 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.769537 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.769567 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.769579 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.769594 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.769608 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.780776 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.793737 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.806418 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.819461 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.831934 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.848088 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.858406 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.871904 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.871947 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.871961 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.871910 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.871982 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.872139 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.890073 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:05Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.942090 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.942221 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.942244 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942355 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942416 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:13.942399475 +0000 UTC m=+92.391094861 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942493 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942526 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942540 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942574 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:13.942563049 +0000 UTC m=+92.391258445 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942617 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:11:13.9426063 +0000 UTC m=+92.391301696 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.942775 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.942800 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942882 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942916 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:13.942906188 +0000 UTC m=+92.391601584 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.942989 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.943002 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.943009 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:05 crc kubenswrapper[4778]: E0312 13:11:05.943042 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:13.943032071 +0000 UTC m=+92.391727467 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.974014 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.974040 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.974047 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.974060 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:05 crc kubenswrapper[4778]: I0312 13:11:05.974069 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:05Z","lastTransitionTime":"2026-03-12T13:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.043960 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:06 crc kubenswrapper[4778]: E0312 13:11:06.044093 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:06 crc kubenswrapper[4778]: E0312 13:11:06.044150 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:14.04413152 +0000 UTC m=+92.492826916 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.076562 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.076586 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.076594 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.076607 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.076617 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.179568 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.179704 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.179724 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.179746 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.179762 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.252873 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.252907 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.252935 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:06 crc kubenswrapper[4778]: E0312 13:11:06.252987 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:06 crc kubenswrapper[4778]: E0312 13:11:06.253120 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.253156 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:06 crc kubenswrapper[4778]: E0312 13:11:06.253263 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:06 crc kubenswrapper[4778]: E0312 13:11:06.253355 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.282099 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.282143 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.282155 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.282171 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.282197 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.384205 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.384241 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.384249 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.384262 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.384271 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.487061 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.487108 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.487121 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.487137 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.487148 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.589985 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.590452 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.590466 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.590489 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.590506 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.662627 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.662911 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.662970 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.662984 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.666746 4778 generic.go:334] "Generic (PLEG): container finished" podID="8a1f8eaa-ac07-4478-be5d-0742de6b43c4" containerID="1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb" exitCode=0 Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.666788 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerDied","Data":"1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.683268 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.688072 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.694971 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.695008 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.695018 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.695031 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.695041 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.695486 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.702633 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.713333 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.726944 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.738539 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.750805 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.765622 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.779256 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.795883 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.796940 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.796975 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.796988 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.797003 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.797013 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.807492 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.816503 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.828549 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.841630 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.851335 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.862822 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.874276 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.886206 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.897893 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.899775 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.899804 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.899816 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.899834 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.899844 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:06Z","lastTransitionTime":"2026-03-12T13:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.908698 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.920075 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.933383 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.943071 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.952581 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.961320 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.969934 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.981959 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:06 crc kubenswrapper[4778]: I0312 13:11:06.993017 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:06Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.001620 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.001652 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.001662 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.001675 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.001683 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.009827 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.022601 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.036628 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.047600 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.081331 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.104598 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.104625 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.104634 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.104649 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.104658 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.208585 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.208663 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.208686 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.208711 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.208728 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.311348 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.311401 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.311417 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.311437 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.311450 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.413782 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.413819 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.413832 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.413848 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.413860 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.515810 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.515840 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.515850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.515863 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.515871 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.625537 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.625597 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.625607 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.625628 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.625640 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.663399 4778 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.675667 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" event={"ID":"8a1f8eaa-ac07-4478-be5d-0742de6b43c4","Type":"ContainerStarted","Data":"f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.699872 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.718264 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.728141 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.728201 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.728226 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.728242 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.728253 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.738334 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.758296 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.777494 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.791828 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.810661 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.821947 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.830662 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.830714 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.830733 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.830786 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.830809 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.836745 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.848857 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.861549 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.873669 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.884387 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.899406 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.920748 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.933090 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.933122 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.933129 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.933142 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.933152 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:07Z","lastTransitionTime":"2026-03-12T13:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:07 crc kubenswrapper[4778]: I0312 13:11:07.936358 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:07Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.035386 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.035441 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.035452 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.035467 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.035476 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.138421 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.138466 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.138480 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.138499 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.138512 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.241149 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.241196 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.241206 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.241218 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.241227 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.253552 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.253552 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.253605 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:08 crc kubenswrapper[4778]: E0312 13:11:08.253710 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.253751 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:08 crc kubenswrapper[4778]: E0312 13:11:08.253842 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:08 crc kubenswrapper[4778]: E0312 13:11:08.253935 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:08 crc kubenswrapper[4778]: E0312 13:11:08.254040 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.343383 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.343419 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.343428 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.343466 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.343482 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.445954 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.445980 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.445987 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.446000 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.446008 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.548812 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.548863 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.548878 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.548899 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.548917 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.651196 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.651237 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.651249 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.651268 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.651279 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.682475 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/0.log" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.687079 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e" exitCode=1 Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.687146 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.688517 4778 scope.go:117] "RemoveContainer" containerID="a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.706402 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.726021 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.752034 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.754314 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.754360 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.754376 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.754398 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.754413 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.767381 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.783498 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.794531 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.805448 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.817540 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.829575 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.847826 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.856028 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.856057 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.856067 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.856080 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.856089 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.862551 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.876386 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.894291 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.906743 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.931476 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:08Z\\\",\\\"message\\\":\\\"from k8s.io/client-go/informers/factory.go:160\\\\nI0312 13:11:08.298033 6399 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 13:11:08.299002 6399 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 13:11:08.299030 6399 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 13:11:08.299060 6399 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 13:11:08.299080 6399 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 13:11:08.299101 6399 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 13:11:08.299110 6399 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 13:11:08.299130 6399 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 13:11:08.299148 6399 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 13:11:08.299157 6399 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 13:11:08.299161 6399 factory.go:656] Stopping watch factory\\\\nI0312 13:11:08.299172 6399 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 13:11:08.299205 6399 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 13:11:08.299247 6399 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.945533 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:08Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.959314 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.959392 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.959413 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.959440 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:08 crc kubenswrapper[4778]: I0312 13:11:08.959457 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:08Z","lastTransitionTime":"2026-03-12T13:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.062426 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.062468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.062480 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.062500 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.062514 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.165400 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.165464 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.165477 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.165494 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.165506 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.281288 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.281325 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.281333 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.281346 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.281355 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.384459 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.384504 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.384515 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.384533 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.384545 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.486057 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.486107 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.486118 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.486133 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.486143 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.587994 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.588045 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.588053 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.588067 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.588076 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.690169 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.690219 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.690232 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.690249 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.690262 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.692295 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/1.log" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.693158 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/0.log" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.696744 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22" exitCode=1 Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.696863 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.696922 4778 scope.go:117] "RemoveContainer" containerID="a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.697404 4778 scope.go:117] "RemoveContainer" containerID="3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22" Mar 12 13:11:09 crc kubenswrapper[4778]: E0312 13:11:09.697629 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.717608 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.735811 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.754713 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5522a1e716ec4cabcf4776dde35c17f9f0a89250cd85474f5a7f94fe8943b1e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:08Z\\\",\\\"message\\\":\\\"from k8s.io/client-go/informers/factory.go:160\\\\nI0312 13:11:08.298033 6399 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 13:11:08.299002 6399 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 13:11:08.299030 6399 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 13:11:08.299060 6399 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 13:11:08.299080 6399 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 13:11:08.299101 6399 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 13:11:08.299110 6399 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 13:11:08.299130 6399 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 13:11:08.299148 6399 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 13:11:08.299157 6399 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 13:11:08.299161 6399 factory.go:656] Stopping watch factory\\\\nI0312 13:11:08.299172 6399 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 13:11:08.299205 6399 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 13:11:08.299247 6399 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:09Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI0312 13:11:09.567525 6707 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0312 13:11:09.567216 6707 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:09.567573 6707 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.765248 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.775014 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.785425 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.792519 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.792551 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.792564 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.792579 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.792590 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.798351 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.813002 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.826500 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.840022 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.849439 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.863162 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.873387 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.885503 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.894102 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.894126 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.894134 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.894148 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.894157 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.896747 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.907775 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.996437 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.996498 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.996517 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.996556 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:09 crc kubenswrapper[4778]: I0312 13:11:09.996592 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:09Z","lastTransitionTime":"2026-03-12T13:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.099662 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.099734 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.099757 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.099785 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.099808 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.201856 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.201883 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.201890 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.201903 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.201912 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.253155 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.253214 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:10 crc kubenswrapper[4778]: E0312 13:11:10.253331 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.253355 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:10 crc kubenswrapper[4778]: E0312 13:11:10.253479 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:10 crc kubenswrapper[4778]: E0312 13:11:10.253578 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.253793 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:10 crc kubenswrapper[4778]: E0312 13:11:10.253912 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.304028 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.304075 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.304086 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.304107 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.304119 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.407272 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.407314 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.407322 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.407350 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.407360 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.510108 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.510170 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.510202 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.510225 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.510240 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.612578 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.612619 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.612631 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.612648 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.612660 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.703350 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/1.log" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.708796 4778 scope.go:117] "RemoveContainer" containerID="3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22" Mar 12 13:11:10 crc kubenswrapper[4778]: E0312 13:11:10.709062 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.714727 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.714797 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.714811 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.714835 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.714850 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.723353 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.740047 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.784836 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.818152 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.818225 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.818241 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.818264 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.818280 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.820758 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.845376 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.856805 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.867035 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.878502 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.888272 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.911813 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.920815 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.920853 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.920861 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.920875 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.920885 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:10Z","lastTransitionTime":"2026-03-12T13:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.934288 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.946135 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.967223 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:10 crc kubenswrapper[4778]: I0312 13:11:10.988686 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:10Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.014219 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:09Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI0312 13:11:09.567525 6707 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0312 13:11:09.567216 6707 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:09.567573 6707 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:11Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.024010 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.024045 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.024055 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.024072 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.024083 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.028367 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:11Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.127927 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.127969 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.127979 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.127995 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.128006 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.230122 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.230172 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.230200 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.230218 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.230230 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.333073 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.333143 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.333161 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.333208 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.333225 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.435963 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.436031 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.436057 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.436088 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.436113 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.538971 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.539067 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.539082 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.539113 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.539129 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.642079 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.642119 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.642130 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.642145 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.642158 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.744380 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.744432 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.744442 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.744455 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.744465 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.776739 4778 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.846168 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.846225 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.846237 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.846277 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.846287 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.948571 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.948624 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.948638 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.948658 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:11 crc kubenswrapper[4778]: I0312 13:11:11.948673 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:11Z","lastTransitionTime":"2026-03-12T13:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.051452 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.051515 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.051528 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.051544 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.051555 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.154251 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.154313 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.154331 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.154354 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.154375 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.246771 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.246837 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.246850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.246865 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.246877 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.256672 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.256786 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.256800 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.256891 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.256977 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.257001 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.257239 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.257220 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.267273 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.278127 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.278372 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.278414 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.278430 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.278452 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.278614 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.293378 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.298950 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.302876 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.302942 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.302962 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.302989 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.303006 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.325579 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.326006 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.330713 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.330772 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.330795 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.330825 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.330862 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.349530 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.350274 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.354223 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.354274 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.354293 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.354316 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.354333 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.367972 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.368093 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: E0312 13:11:12.368271 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.370626 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.370693 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.370705 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.370722 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.370735 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.385554 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.396838 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.409992 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.421753 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.435687 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.446544 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.457043 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.471846 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.473610 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.473658 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.473669 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.473684 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.473697 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.485230 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.504019 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:09Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI0312 13:11:09.567525 6707 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0312 13:11:09.567216 6707 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:09.567573 6707 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.513668 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:12Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.576817 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.576891 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.576908 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.576932 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.576952 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.680020 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.680083 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.680101 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.680125 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.680144 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.782454 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.782492 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.782500 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.782513 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.782522 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.885862 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.885918 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.885931 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.885951 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.885964 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.989087 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.989148 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.989162 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.989208 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:12 crc kubenswrapper[4778]: I0312 13:11:12.989234 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:12Z","lastTransitionTime":"2026-03-12T13:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.091850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.091903 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.091918 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.091938 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.091951 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.195205 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.195238 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.195245 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.195257 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.195265 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.297322 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.297384 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.297394 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.297409 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.297420 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.400211 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.400244 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.400255 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.400272 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.400285 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.503298 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.503346 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.503369 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.503393 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.503408 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.606362 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.606418 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.606435 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.606457 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.606472 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.709689 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.709741 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.709752 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.709767 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.709778 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.811702 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.811729 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.811737 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.811749 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.811758 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.914111 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.914155 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.914168 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.914203 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:13 crc kubenswrapper[4778]: I0312 13:11:13.914222 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:13Z","lastTransitionTime":"2026-03-12T13:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.016972 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.017068 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.017101 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.017134 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.017155 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.033432 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033583 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:11:30.03355455 +0000 UTC m=+108.482249946 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.033635 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.033711 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.033731 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.033766 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033822 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033833 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033844 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033886 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033821 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033888 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:30.033880848 +0000 UTC m=+108.482576244 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033843 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033996 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:30.033975301 +0000 UTC m=+108.482670747 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.033996 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.034013 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.034017 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:30.034008722 +0000 UTC m=+108.482704228 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.034045 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:30.034033352 +0000 UTC m=+108.482728818 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.119762 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.119808 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.119819 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.119834 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.119847 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.134666 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.134797 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.134870 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:11:30.134853844 +0000 UTC m=+108.583549240 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.222603 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.222652 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.222662 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.222681 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.222691 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.253143 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.253297 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.253473 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.253481 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.253489 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.253601 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.254248 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.254359 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.254534 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:11:14 crc kubenswrapper[4778]: E0312 13:11:14.254793 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.325583 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.325614 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.325630 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.325646 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.325657 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.428346 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.428415 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.428428 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.428446 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.428462 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.531174 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.531258 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.531270 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.531289 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.531305 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.634630 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.634703 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.634715 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.634729 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.634739 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.737086 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.737157 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.737232 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.737262 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.737282 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.840597 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.840898 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.840918 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.840955 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.840972 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.943085 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.943133 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.943144 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.943161 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:14 crc kubenswrapper[4778]: I0312 13:11:14.943172 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:14Z","lastTransitionTime":"2026-03-12T13:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.045840 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.045885 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.045898 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.045914 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.045927 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.148045 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.148107 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.148119 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.148134 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.148144 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.250293 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.250353 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.250368 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.250391 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.250407 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.353096 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.353153 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.353165 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.353198 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.353212 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.456332 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.456383 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.456400 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.456419 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.456432 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.562367 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.562427 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.562441 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.562462 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.562479 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.665249 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.665291 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.665303 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.665319 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.665331 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.768730 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.768816 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.768831 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.768862 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.768944 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.871009 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.871050 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.871058 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.871074 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.871085 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.974064 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.974116 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.974126 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.974146 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:15 crc kubenswrapper[4778]: I0312 13:11:15.974158 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:15Z","lastTransitionTime":"2026-03-12T13:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.076566 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.076611 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.076622 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.076638 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.076649 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.178936 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.178976 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.178986 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.179000 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.179009 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.253113 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.253144 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.253113 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.253169 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:16 crc kubenswrapper[4778]: E0312 13:11:16.253280 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:16 crc kubenswrapper[4778]: E0312 13:11:16.253497 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:16 crc kubenswrapper[4778]: E0312 13:11:16.253521 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:16 crc kubenswrapper[4778]: E0312 13:11:16.253587 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.281713 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.281752 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.281763 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.281780 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.281793 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.385303 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.385360 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.385376 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.385398 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.385414 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.488872 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.488932 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.488947 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.488969 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.488982 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.591966 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.592012 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.592024 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.592041 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.592056 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.694747 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.694800 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.694818 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.694842 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.694859 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.797755 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.797816 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.797834 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.797859 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.797877 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.900323 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.900378 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.900394 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.900414 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:16 crc kubenswrapper[4778]: I0312 13:11:16.900430 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:16Z","lastTransitionTime":"2026-03-12T13:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.002796 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.002853 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.002870 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.002892 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.002910 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.105083 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.105114 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.105123 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.105136 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.105145 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.207492 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.207546 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.207563 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.207581 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.207594 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.310756 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.310797 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.310809 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.310824 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.310833 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.413427 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.413486 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.413505 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.413532 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.413557 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.516602 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.516648 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.516663 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.516681 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.516693 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.619258 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.619314 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.619323 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.619337 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.619345 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.722216 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.722284 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.722303 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.722326 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.722348 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.824830 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.824903 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.824920 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.824944 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.824960 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.928058 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.928109 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.928121 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.928144 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:17 crc kubenswrapper[4778]: I0312 13:11:17.928156 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:17Z","lastTransitionTime":"2026-03-12T13:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.035080 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.035154 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.035169 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.035220 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.035243 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.138757 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.138813 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.138827 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.138845 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.138858 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.241840 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.241915 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.241930 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.241951 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.241964 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.253361 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.253398 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.253446 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.253373 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:18 crc kubenswrapper[4778]: E0312 13:11:18.253607 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:18 crc kubenswrapper[4778]: E0312 13:11:18.253714 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:18 crc kubenswrapper[4778]: E0312 13:11:18.253803 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:18 crc kubenswrapper[4778]: E0312 13:11:18.253928 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.345598 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.345639 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.345649 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.345666 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.345677 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.449760 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.450383 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.450423 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.450451 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.450468 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.553777 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.554242 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.554339 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.554500 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.554608 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.657754 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.657821 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.657838 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.657861 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.657877 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.760579 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.760644 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.760653 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.760671 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.760681 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.862626 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.862659 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.862682 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.862696 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.862704 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.966384 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.966479 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.966544 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.966571 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:18 crc kubenswrapper[4778]: I0312 13:11:18.966637 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:18Z","lastTransitionTime":"2026-03-12T13:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.069798 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.070367 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.070463 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.070588 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.070687 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.174774 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.174830 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.174842 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.174866 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.174880 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.278780 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.278828 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.278842 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.278863 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.278880 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.408980 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.409050 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.409064 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.409092 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.409104 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.511942 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.511995 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.512030 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.512050 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.512063 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.615021 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.615080 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.615117 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.615136 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.615149 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.718330 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.718365 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.718376 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.718395 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.718414 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.821289 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.821326 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.821337 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.821353 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.821365 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.923867 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.923909 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.923920 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.923935 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:19 crc kubenswrapper[4778]: I0312 13:11:19.923945 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:19Z","lastTransitionTime":"2026-03-12T13:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.026319 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.026372 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.026382 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.026396 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.026406 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.128824 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.128865 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.128873 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.128887 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.128898 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.230901 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.230959 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.230974 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.230992 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.231004 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.253335 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.253434 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:20 crc kubenswrapper[4778]: E0312 13:11:20.253497 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.253335 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.253345 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:20 crc kubenswrapper[4778]: E0312 13:11:20.253688 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:20 crc kubenswrapper[4778]: E0312 13:11:20.253777 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:20 crc kubenswrapper[4778]: E0312 13:11:20.253584 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.333610 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.333659 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.333673 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.333691 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.333702 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.437046 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.437107 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.437130 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.437161 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.437217 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.539553 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.539689 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.539709 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.539732 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.539749 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.642334 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.642362 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.642369 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.642381 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.642389 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.744418 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.744486 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.744524 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.744541 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.744550 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.847119 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.847356 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.847397 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.847426 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.847448 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.950430 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.950477 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.950488 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.950505 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:20 crc kubenswrapper[4778]: I0312 13:11:20.950517 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:20Z","lastTransitionTime":"2026-03-12T13:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.052426 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.052476 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.052489 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.052506 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.052518 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.155143 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.155201 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.155214 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.155228 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.155240 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.254825 4778 scope.go:117] "RemoveContainer" containerID="3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.258076 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.258120 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.258132 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.258147 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.258160 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.361098 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.361172 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.361286 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.361315 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.361328 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.463536 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.463579 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.463588 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.463604 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.463614 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.566010 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.566073 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.566091 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.566111 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.566125 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.668850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.668899 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.668915 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.668935 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.668950 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.785866 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.785904 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.785917 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.785932 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.785968 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.791498 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/1.log" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.794026 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.796728 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.808145 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.826340 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.841574 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.852068 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.860929 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.874073 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.887156 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.888206 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.888333 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.888442 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.888584 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.888691 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.902668 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.914499 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.929490 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.948433 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.961446 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.975783 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.991493 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.991552 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.991565 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.991581 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.991593 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:21Z","lastTransitionTime":"2026-03-12T13:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:21 crc kubenswrapper[4778]: I0312 13:11:21.994050 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:09Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI0312 13:11:09.567525 6707 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0312 13:11:09.567216 6707 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:09.567573 6707 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:21Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.005304 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.026284 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.093908 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.093980 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.093996 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.094046 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.094065 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.195955 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.195996 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.196004 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.196022 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.196032 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.253276 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.253402 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.253436 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.253466 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.253541 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.253439 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.253612 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.253649 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.268613 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.289814 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.299657 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.299738 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.299762 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.299793 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.299816 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.304348 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.320018 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.337780 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.350884 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.370612 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.386822 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.402217 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.402270 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.402306 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.402329 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.402345 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.406359 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:09Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI0312 13:11:09.567525 6707 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0312 13:11:09.567216 6707 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:09.567573 6707 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.420476 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.434819 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.447253 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.463767 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.480392 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.498262 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.504698 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.504746 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.504759 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.504779 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.504791 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.517672 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.538290 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.538343 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.538353 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.538371 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.538388 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.554207 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.559611 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.559673 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.559690 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.559712 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.559726 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.574221 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.578166 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.578223 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.578241 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.578262 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.578276 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.590875 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.594827 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.594879 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.594891 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.594910 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.594924 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.609176 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.613984 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.614037 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.614051 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.614107 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.614123 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.633447 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.633617 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.635754 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.635819 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.635832 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.635849 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.635859 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.738817 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.738882 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.738902 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.738927 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.738945 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.801177 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/2.log" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.802213 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/1.log" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.805813 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824" exitCode=1 Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.805875 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.806270 4778 scope.go:117] "RemoveContainer" containerID="3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.806744 4778 scope.go:117] "RemoveContainer" containerID="cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824" Mar 12 13:11:22 crc kubenswrapper[4778]: E0312 13:11:22.806975 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.821961 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.839700 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.843228 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.843299 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.843311 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.843331 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.843381 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.854420 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.868101 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.881942 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.898424 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.918728 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.930533 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.945466 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.946367 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.946436 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.946457 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.946482 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.946502 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:22Z","lastTransitionTime":"2026-03-12T13:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.964913 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa32715eae6ff23b04c9b3865505ff2ed911d459033b9d6912866b5df2f8d22\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:09Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI0312 13:11:09.567525 6707 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0312 13:11:09.567216 6707 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:09Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:09.567573 6707 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.977853 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:22 crc kubenswrapper[4778]: I0312 13:11:22.994717 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.009512 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.025859 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.038653 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.049272 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.049692 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.049786 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.049823 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.049865 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.049913 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.154585 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.154692 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.154710 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.154736 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.154765 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.257620 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.257687 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.257701 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.257724 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.257739 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.361281 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.361361 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.361376 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.361393 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.361406 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.464161 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.464288 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.464311 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.464341 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.464363 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.566700 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.566768 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.566778 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.566792 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.566802 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.669140 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.669231 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.669256 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.669283 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.669308 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.772217 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.772283 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.772309 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.772333 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.772379 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.810510 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/2.log" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.813493 4778 scope.go:117] "RemoveContainer" containerID="cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824" Mar 12 13:11:23 crc kubenswrapper[4778]: E0312 13:11:23.813633 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.826469 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.835911 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.845365 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.857327 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.867094 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.874231 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.874306 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.874320 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.874338 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.874352 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.880990 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.898367 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.910094 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.925580 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.939448 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.950494 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.960435 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.970156 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.976989 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.977052 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.977064 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.977078 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.977087 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:23Z","lastTransitionTime":"2026-03-12T13:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.981060 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:23 crc kubenswrapper[4778]: I0312 13:11:23.994676 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:23Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.008879 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:24Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.079407 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.079445 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.079458 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.079473 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.079483 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.183928 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.183985 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.184002 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.184023 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.184041 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.253082 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.253152 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.253168 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:24 crc kubenswrapper[4778]: E0312 13:11:24.253282 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.253300 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:24 crc kubenswrapper[4778]: E0312 13:11:24.253408 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:24 crc kubenswrapper[4778]: E0312 13:11:24.253516 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:24 crc kubenswrapper[4778]: E0312 13:11:24.254066 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.267636 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.287236 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.287303 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.287316 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.287333 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.287346 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.390003 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.390051 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.390063 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.390077 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.390088 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.492327 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.492367 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.492376 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.492389 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.492399 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.595370 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.595410 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.595421 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.595436 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.595447 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.698213 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.698280 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.698304 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.698340 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.698362 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.801002 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.801050 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.801063 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.801078 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.801089 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.904026 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.904307 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.904367 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.904435 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:24 crc kubenswrapper[4778]: I0312 13:11:24.904550 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:24Z","lastTransitionTime":"2026-03-12T13:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.008046 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.008116 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.008137 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.008160 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.008177 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.110979 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.111324 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.111476 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.111669 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.111857 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.214712 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.214768 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.214785 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.214807 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.214825 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.254097 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.320523 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.320588 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.320601 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.320619 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.320631 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.423230 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.423280 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.423291 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.423310 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.423324 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.526637 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.526679 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.526691 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.526709 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.526721 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.634634 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.634668 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.634682 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.634702 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.634717 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.737984 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.738047 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.738068 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.738095 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.738117 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.822853 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.826827 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.827678 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.840381 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.840419 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.840429 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.840445 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.840456 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.845706 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.859242 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.869941 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.879099 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.890852 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.913780 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.928630 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.939279 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.942643 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.942674 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.942682 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.942695 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.942704 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:25Z","lastTransitionTime":"2026-03-12T13:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.951705 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.965559 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.979490 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:25 crc kubenswrapper[4778]: I0312 13:11:25.992219 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:25Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.005204 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:26Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.019668 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:26Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.038740 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:26Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.044737 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.044766 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.044775 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.044832 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.044842 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.049684 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:26Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.062465 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:26Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.147920 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.147976 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.147993 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.148018 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.148032 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.250242 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.250297 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.250315 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.250339 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.250355 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.252806 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.252841 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.252972 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:26 crc kubenswrapper[4778]: E0312 13:11:26.253063 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.253082 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:26 crc kubenswrapper[4778]: E0312 13:11:26.253245 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:26 crc kubenswrapper[4778]: E0312 13:11:26.253451 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:26 crc kubenswrapper[4778]: E0312 13:11:26.253612 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.352839 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.352874 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.352886 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.352899 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.352907 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.455142 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.455170 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.455191 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.455204 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.455214 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.558039 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.558087 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.558100 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.558117 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.558129 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.660702 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.660754 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.660767 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.660785 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.660798 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.763797 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.763850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.763864 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.763894 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.763911 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.866901 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.866943 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.866956 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.866973 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.866984 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.969045 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.969086 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.969099 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.969116 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:26 crc kubenswrapper[4778]: I0312 13:11:26.969130 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:26Z","lastTransitionTime":"2026-03-12T13:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.071607 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.071648 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.071659 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.071674 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.071685 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.174443 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.174497 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.174506 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.174519 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.174530 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.277322 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.277824 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.277889 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.277986 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.278059 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.380622 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.380831 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.380926 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.381004 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.381084 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.483495 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.483547 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.483564 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.483589 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.483607 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.586298 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.586373 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.586394 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.586427 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.586450 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.689616 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.690035 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.690366 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.690577 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.690772 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.793472 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.793744 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.793843 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.793941 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.794027 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.897955 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.898022 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.898041 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.898065 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:27 crc kubenswrapper[4778]: I0312 13:11:27.898083 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:27Z","lastTransitionTime":"2026-03-12T13:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.000404 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.000453 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.000464 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.000480 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.000491 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.103796 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.103853 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.103871 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.103895 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.103916 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.206739 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.207099 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.207205 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.207399 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.207591 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.252911 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.252912 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.252998 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.253063 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:28 crc kubenswrapper[4778]: E0312 13:11:28.253247 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:28 crc kubenswrapper[4778]: E0312 13:11:28.253430 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:28 crc kubenswrapper[4778]: E0312 13:11:28.253573 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:28 crc kubenswrapper[4778]: E0312 13:11:28.253740 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.310218 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.310276 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.310294 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.310316 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.310328 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.413823 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.413869 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.413883 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.413902 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.413918 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.516675 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.516721 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.516733 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.516749 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.516762 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.619211 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.619259 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.619273 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.619291 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.619302 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.721687 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.721727 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.721739 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.721755 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.721767 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.824074 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.824131 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.824147 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.824170 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.824212 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.926541 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.926619 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.926634 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.926652 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:28 crc kubenswrapper[4778]: I0312 13:11:28.926692 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:28Z","lastTransitionTime":"2026-03-12T13:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.029371 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.029604 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.029704 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.029774 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.029830 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.132432 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.132476 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.132488 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.132504 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.132515 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.235706 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.235768 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.235780 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.235798 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.235809 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.337854 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.337908 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.337919 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.337937 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.337950 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.440082 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.440402 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.440474 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.440747 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.440806 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.543199 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.543246 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.543258 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.543285 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.543330 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.646581 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.646616 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.646628 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.646646 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.646659 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.749771 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.749832 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.749850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.749872 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.749889 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.852950 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.853013 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.853025 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.853046 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.853059 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.955520 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.955564 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.955575 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.955590 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:29 crc kubenswrapper[4778]: I0312 13:11:29.955601 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:29Z","lastTransitionTime":"2026-03-12T13:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.057990 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.058051 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.058067 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.058087 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.058102 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.101096 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.101340 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101392 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:02.101347647 +0000 UTC m=+140.550043073 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.101537 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101593 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101623 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101634 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101657 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101670 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101678 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101686 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:02.101670376 +0000 UTC m=+140.550365822 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101693 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101709 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:02.101700487 +0000 UTC m=+140.550395883 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.101597 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101750 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:02.101735707 +0000 UTC m=+140.550431293 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.101773 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101837 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.101861 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:02.101854081 +0000 UTC m=+140.550549477 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.160680 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.160729 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.160740 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.160758 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.160771 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.202797 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.203030 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.203137 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:02.203113344 +0000 UTC m=+140.651808780 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.253803 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.253842 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.253873 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.253934 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.253800 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.254119 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.254595 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:30 crc kubenswrapper[4778]: E0312 13:11:30.254761 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.262125 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.262228 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.262255 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.262282 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.262302 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.364753 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.364805 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.364814 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.364827 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.364835 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.466803 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.467160 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.467322 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.467442 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.467531 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.570289 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.570353 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.570372 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.570395 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.570413 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.673147 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.673217 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.673228 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.673246 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.673257 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.775936 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.775981 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.775994 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.776010 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.776022 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.877926 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.877962 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.877973 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.877988 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.878013 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.981411 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.981455 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.981467 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.981484 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:30 crc kubenswrapper[4778]: I0312 13:11:30.981495 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:30Z","lastTransitionTime":"2026-03-12T13:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.083773 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.083821 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.083841 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.083859 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.083872 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.186783 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.187098 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.187111 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.187149 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.187164 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.289997 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.290043 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.290056 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.290074 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.290085 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.392988 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.393021 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.393032 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.393045 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.393054 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.495762 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.495806 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.495817 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.495835 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.495847 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.598494 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.598548 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.598558 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.598576 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.598585 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.701572 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.701626 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.701644 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.701663 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.701724 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.807672 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.807716 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.807726 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.807742 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.807752 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.910831 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.910896 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.910908 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.910924 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:31 crc kubenswrapper[4778]: I0312 13:11:31.910936 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:31Z","lastTransitionTime":"2026-03-12T13:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.014204 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.014251 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.014261 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.014278 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.014289 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.117149 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.117522 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.117679 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.117853 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.118041 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.220753 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.220796 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.220808 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.220824 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.220836 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.253293 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.253435 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.253722 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.253822 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.253924 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.254004 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.254020 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.254233 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.267059 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.289565 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.302324 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.315108 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.323526 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.323743 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.323859 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.323949 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.324036 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.328741 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.340920 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.356129 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.370555 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.381820 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.394154 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.413499 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.425574 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.426960 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.426977 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.426986 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.426999 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.427009 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.447323 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.465145 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.480943 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.493324 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.505460 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.529817 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.529874 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.529884 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.529905 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.529919 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.632524 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.632575 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.632587 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.632606 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.632639 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.735547 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.735602 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.735614 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.735634 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.735651 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.838446 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.838545 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.838566 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.838590 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.838607 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.862983 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.863046 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.863064 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.863090 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.863113 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.881365 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.886449 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.886514 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.886528 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.886543 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.886554 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.899585 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.904932 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.905020 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.905461 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.905545 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.905823 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.925598 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.930418 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.930483 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.930499 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.930518 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.930810 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.946929 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.951653 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.951714 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.951735 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.951760 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.951779 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.974267 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:32Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:32 crc kubenswrapper[4778]: E0312 13:11:32.974685 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.976563 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.976616 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.976627 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.976647 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:32 crc kubenswrapper[4778]: I0312 13:11:32.976659 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:32Z","lastTransitionTime":"2026-03-12T13:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.079423 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.079503 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.079522 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.079545 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.079563 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.183218 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.183288 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.183302 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.183322 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.183334 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.286408 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.286457 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.286468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.286486 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.286498 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.389478 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.389754 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.389835 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.389948 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.390243 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.493310 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.493347 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.493358 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.493375 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.493386 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.596192 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.596447 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.596510 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.596571 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.596634 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.699331 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.699371 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.699382 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.699396 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.699406 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.801625 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.801671 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.801686 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.801706 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.801721 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.905012 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.905246 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.905315 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.905375 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:33 crc kubenswrapper[4778]: I0312 13:11:33.905459 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:33Z","lastTransitionTime":"2026-03-12T13:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.008201 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.008473 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.008562 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.008661 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.008738 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.111514 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.111565 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.111582 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.111603 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.111619 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.214372 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.214441 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.214462 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.214489 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.214512 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.253706 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.253776 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:34 crc kubenswrapper[4778]: E0312 13:11:34.253845 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.253879 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.253706 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:34 crc kubenswrapper[4778]: E0312 13:11:34.253926 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:34 crc kubenswrapper[4778]: E0312 13:11:34.254001 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:34 crc kubenswrapper[4778]: E0312 13:11:34.254054 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.318340 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.318397 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.318413 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.318435 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.318446 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.420833 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.420888 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.420905 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.420928 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.420945 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.524317 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.524368 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.524380 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.524399 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.524410 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.627913 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.627964 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.627976 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.627992 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.628005 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.731090 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.731152 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.731162 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.731177 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.731208 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.833986 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.834093 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.834103 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.834117 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.834125 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.936798 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.936831 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.936841 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.936854 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:34 crc kubenswrapper[4778]: I0312 13:11:34.936863 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:34Z","lastTransitionTime":"2026-03-12T13:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.039156 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.039236 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.039248 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.039263 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.039272 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.149699 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.149746 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.149759 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.149776 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.149787 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.252251 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.252288 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.252297 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.252310 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.252319 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.354644 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.354714 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.354733 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.354759 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.354777 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.457237 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.457289 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.457301 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.457318 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.457330 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.559664 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.559745 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.559773 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.559803 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.559826 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.661997 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.662232 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.662313 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.662429 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.662527 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.765431 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.765497 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.765520 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.765548 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.765569 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.868837 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.868898 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.868915 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.868934 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.868947 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.972028 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.972073 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.972090 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.972105 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:35 crc kubenswrapper[4778]: I0312 13:11:35.972114 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:35Z","lastTransitionTime":"2026-03-12T13:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.074354 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.074389 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.074401 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.074419 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.074431 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.177851 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.177931 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.177954 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.177986 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.178010 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.253010 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:36 crc kubenswrapper[4778]: E0312 13:11:36.253452 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.253124 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:36 crc kubenswrapper[4778]: E0312 13:11:36.253675 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.253080 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:36 crc kubenswrapper[4778]: E0312 13:11:36.253842 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.253128 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:36 crc kubenswrapper[4778]: E0312 13:11:36.254021 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.280518 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.280741 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.280800 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.280859 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.280911 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.383164 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.383238 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.383249 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.383265 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.383277 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.485836 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.485895 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.485907 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.485925 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.485937 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.589905 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.589970 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.589981 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.589997 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.590028 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.692802 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.692865 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.692882 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.692907 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.692924 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.795652 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.795697 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.795710 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.795726 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.795735 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.898657 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.898702 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.898720 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.898744 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:36 crc kubenswrapper[4778]: I0312 13:11:36.898764 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:36Z","lastTransitionTime":"2026-03-12T13:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.001467 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.001541 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.001553 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.001571 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.001582 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.103792 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.103838 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.103848 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.103860 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.103869 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.206095 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.206176 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.206224 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.206240 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.206254 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.254092 4778 scope.go:117] "RemoveContainer" containerID="cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824" Mar 12 13:11:37 crc kubenswrapper[4778]: E0312 13:11:37.254297 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.308757 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.308998 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.309081 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.309148 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.309228 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.411385 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.411468 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.411484 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.411507 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.411523 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.514931 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.516001 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.516344 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.516454 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.516556 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.618843 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.619450 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.619521 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.619630 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.619699 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.722619 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.722679 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.722692 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.722710 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.722722 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.825433 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.825484 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.825495 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.825511 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.825522 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.927972 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.928046 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.928059 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.928100 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:37 crc kubenswrapper[4778]: I0312 13:11:37.928126 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:37Z","lastTransitionTime":"2026-03-12T13:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.030504 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.030547 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.030559 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.030574 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.030586 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.132974 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.133033 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.133042 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.133057 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.133067 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.235308 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.235345 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.235357 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.235371 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.235382 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.253742 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.253771 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.253774 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.253787 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:38 crc kubenswrapper[4778]: E0312 13:11:38.253859 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:38 crc kubenswrapper[4778]: E0312 13:11:38.254000 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:38 crc kubenswrapper[4778]: E0312 13:11:38.254066 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:38 crc kubenswrapper[4778]: E0312 13:11:38.254215 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.337328 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.337365 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.337376 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.337390 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.337401 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.439663 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.439704 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.439714 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.439730 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.439745 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.542687 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.542752 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.542772 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.542799 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.542820 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.645035 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.645610 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.645705 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.645790 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.645884 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.749348 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.749403 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.749418 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.749437 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.749457 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.852403 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.852669 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.852737 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.852815 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.852875 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.955649 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.955731 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.955744 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.955766 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:38 crc kubenswrapper[4778]: I0312 13:11:38.955780 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:38Z","lastTransitionTime":"2026-03-12T13:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.057755 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.057784 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.057792 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.057805 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.057816 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.160809 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.160881 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.160902 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.160930 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.160953 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.263824 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.263858 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.263869 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.263884 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.263895 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.372342 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.372386 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.372395 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.372409 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.372419 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.474477 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.474529 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.474547 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.474565 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.474578 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.576866 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.576959 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.576982 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.577012 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.577033 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.679537 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.679586 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.679599 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.679616 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.679631 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.782048 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.782089 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.782100 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.782115 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.782127 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.884591 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.884638 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.884646 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.884663 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.884673 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.987749 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.987816 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.987832 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.987865 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:39 crc kubenswrapper[4778]: I0312 13:11:39.987879 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:39Z","lastTransitionTime":"2026-03-12T13:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.090386 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.090430 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.090443 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.090460 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.090472 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.192869 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.192911 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.192923 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.192939 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.192948 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.252929 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.253049 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.253135 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:40 crc kubenswrapper[4778]: E0312 13:11:40.253083 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:40 crc kubenswrapper[4778]: E0312 13:11:40.253327 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:40 crc kubenswrapper[4778]: E0312 13:11:40.253445 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.253718 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:40 crc kubenswrapper[4778]: E0312 13:11:40.253787 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.295940 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.295976 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.295985 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.296000 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.296010 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.399886 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.399953 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.399976 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.400003 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.400024 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.502932 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.502958 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.502966 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.502979 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.502988 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.606106 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.606170 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.606218 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.606247 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.606265 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.709604 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.709641 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.709652 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.709668 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.709679 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.812160 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.812223 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.812232 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.812247 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.812257 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.915628 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.915660 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.915670 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.915683 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:40 crc kubenswrapper[4778]: I0312 13:11:40.915692 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:40Z","lastTransitionTime":"2026-03-12T13:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.018116 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.018144 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.018152 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.018165 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.018176 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.120365 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.120412 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.120428 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.120453 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.120466 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.222783 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.222823 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.222833 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.222850 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.222860 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.325234 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.325294 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.325309 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.325329 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.325345 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.429239 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.429290 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.429301 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.429318 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.429329 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.532558 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.532622 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.532635 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.532661 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.532675 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.637482 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.637541 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.637557 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.637578 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.637592 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.739691 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.739736 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.739749 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.739765 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.739777 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.842511 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.842556 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.842566 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.842580 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.842590 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.944703 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.944744 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.944754 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.944769 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:41 crc kubenswrapper[4778]: I0312 13:11:41.944779 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:41Z","lastTransitionTime":"2026-03-12T13:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.047388 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.047426 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.047434 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.047447 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.047456 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:42Z","lastTransitionTime":"2026-03-12T13:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.149733 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.149774 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.149783 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.149797 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.149806 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:42Z","lastTransitionTime":"2026-03-12T13:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:42 crc kubenswrapper[4778]: E0312 13:11:42.250927 4778 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.253396 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.253425 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.253455 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.253448 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:42 crc kubenswrapper[4778]: E0312 13:11:42.253658 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:42 crc kubenswrapper[4778]: E0312 13:11:42.253761 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:42 crc kubenswrapper[4778]: E0312 13:11:42.253838 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:42 crc kubenswrapper[4778]: E0312 13:11:42.253890 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.278924 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.293519 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.304841 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.316880 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.332659 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.347454 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.358357 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: E0312 13:11:42.368061 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.373980 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.385952 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.394444 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.408298 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.423492 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.443572 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.457627 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.468048 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.477155 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:42 crc kubenswrapper[4778]: I0312 13:11:42.488229 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:42Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.323661 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.323742 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.323756 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.323779 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.323801 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:43Z","lastTransitionTime":"2026-03-12T13:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:43 crc kubenswrapper[4778]: E0312 13:11:43.339597 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.343241 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.343369 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.343401 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.343426 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.343444 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:43Z","lastTransitionTime":"2026-03-12T13:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:43 crc kubenswrapper[4778]: E0312 13:11:43.360701 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.365508 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.365563 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.365583 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.365606 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.365636 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:43Z","lastTransitionTime":"2026-03-12T13:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:43 crc kubenswrapper[4778]: E0312 13:11:43.384968 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.396962 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.397020 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.397032 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.397050 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.397063 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:43Z","lastTransitionTime":"2026-03-12T13:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:43 crc kubenswrapper[4778]: E0312 13:11:43.415530 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.419338 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.419379 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.419389 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.419417 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.419426 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:43Z","lastTransitionTime":"2026-03-12T13:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:43 crc kubenswrapper[4778]: E0312 13:11:43.432964 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: E0312 13:11:43.433173 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.542534 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.559985 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.572887 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.587937 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.601465 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.613149 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.627172 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.639178 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.661685 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.673590 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.685324 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.697215 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.709179 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.721923 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.734367 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.745600 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.755545 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:43 crc kubenswrapper[4778]: I0312 13:11:43.774294 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:43Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:44 crc kubenswrapper[4778]: I0312 13:11:44.252993 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:44 crc kubenswrapper[4778]: I0312 13:11:44.253104 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:44 crc kubenswrapper[4778]: E0312 13:11:44.253177 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:44 crc kubenswrapper[4778]: E0312 13:11:44.253385 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:44 crc kubenswrapper[4778]: I0312 13:11:44.253456 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:44 crc kubenswrapper[4778]: I0312 13:11:44.253483 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:44 crc kubenswrapper[4778]: E0312 13:11:44.253544 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:44 crc kubenswrapper[4778]: E0312 13:11:44.253634 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:46 crc kubenswrapper[4778]: I0312 13:11:46.253264 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:46 crc kubenswrapper[4778]: I0312 13:11:46.253369 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:46 crc kubenswrapper[4778]: I0312 13:11:46.253436 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:46 crc kubenswrapper[4778]: I0312 13:11:46.253489 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:46 crc kubenswrapper[4778]: E0312 13:11:46.253894 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:46 crc kubenswrapper[4778]: E0312 13:11:46.254312 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:46 crc kubenswrapper[4778]: E0312 13:11:46.254431 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:46 crc kubenswrapper[4778]: E0312 13:11:46.254530 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:47 crc kubenswrapper[4778]: E0312 13:11:47.369378 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.898990 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fhcz6_1e7037a8-a966-4df0-9f94-fe2dd3e2de6e/kube-multus/0.log" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.899050 4778 generic.go:334] "Generic (PLEG): container finished" podID="1e7037a8-a966-4df0-9f94-fe2dd3e2de6e" containerID="5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac" exitCode=1 Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.899083 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fhcz6" event={"ID":"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e","Type":"ContainerDied","Data":"5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac"} Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.899479 4778 scope.go:117] "RemoveContainer" containerID="5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.921984 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:47Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.938341 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:47Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.952004 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:47Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.966947 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:47Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.977430 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:47Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:47 crc kubenswrapper[4778]: I0312 13:11:47.990837 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:47Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.009620 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.021098 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.035961 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.050431 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.066766 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.080663 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.095628 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.108313 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.132728 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.149083 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.162721 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.256569 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:48 crc kubenswrapper[4778]: E0312 13:11:48.256742 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.256996 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:48 crc kubenswrapper[4778]: E0312 13:11:48.257117 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.257945 4778 scope.go:117] "RemoveContainer" containerID="cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.258315 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:48 crc kubenswrapper[4778]: E0312 13:11:48.258380 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.258514 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:48 crc kubenswrapper[4778]: E0312 13:11:48.258588 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.904740 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/2.log" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.907993 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.908949 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.912493 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fhcz6_1e7037a8-a966-4df0-9f94-fe2dd3e2de6e/kube-multus/0.log" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.912529 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fhcz6" event={"ID":"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e","Type":"ContainerStarted","Data":"44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e"} Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.924399 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.943123 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.954120 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.971176 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:48 crc kubenswrapper[4778]: I0312 13:11:48.985957 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.001658 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:48Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.014841 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.030084 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.044408 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.068762 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.086644 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.098246 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.114429 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.129047 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.144840 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.158936 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.171816 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.203133 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.224727 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.240864 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.255216 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.271884 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.286918 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.301254 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.313942 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.323647 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.333372 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.345761 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.364401 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.388692 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.402409 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.416221 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.430854 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.445801 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.918713 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/3.log" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.919359 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/2.log" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.923041 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" exitCode=1 Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.923125 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.923231 4778 scope.go:117] "RemoveContainer" containerID="cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.924061 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:11:49 crc kubenswrapper[4778]: E0312 13:11:49.924341 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.941371 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.953820 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.964574 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.976778 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:49 crc kubenswrapper[4778]: I0312 13:11:49.986251 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.006973 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.018429 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.032383 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.048687 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.059858 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.070639 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.083428 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.095520 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.109724 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.132310 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc5950d10319c346220cc78cf45052b31ea05a32f6d5f2511a963110c4a17824\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:22Z\\\",\\\"message\\\":\\\"ificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z\\\\nI0312 13:11:22.188287 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF0312 13:11:22.188260 6941 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:22Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:22.188294 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j\\\\nI0312 13:11:22.188297 6941 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4dfhs\\\\nI031\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:49Z\\\",\\\"message\\\":\\\" network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:49.218951 7264 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f32857b5-f652-4313-a0d7-455c3156dd99\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.146550 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.161767 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.253733 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.253804 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.253819 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.253805 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:50 crc kubenswrapper[4778]: E0312 13:11:50.253978 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:50 crc kubenswrapper[4778]: E0312 13:11:50.254032 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:50 crc kubenswrapper[4778]: E0312 13:11:50.254092 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:50 crc kubenswrapper[4778]: E0312 13:11:50.254146 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.929593 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/3.log" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.934550 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:11:50 crc kubenswrapper[4778]: E0312 13:11:50.934806 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.949984 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.966948 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:50 crc kubenswrapper[4778]: I0312 13:11:50.989626 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:50Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.020209 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:49Z\\\",\\\"message\\\":\\\" network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:49.218951 7264 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f32857b5-f652-4313-a0d7-455c3156dd99\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.033297 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.043874 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.053387 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.064387 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.080815 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.091303 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.101263 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.109948 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.120135 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.131996 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.143670 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.154360 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:51 crc kubenswrapper[4778]: I0312 13:11:51.163246 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:51Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.253964 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:52 crc kubenswrapper[4778]: E0312 13:11:52.254091 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.254360 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:52 crc kubenswrapper[4778]: E0312 13:11:52.254417 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.254642 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.254694 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:52 crc kubenswrapper[4778]: E0312 13:11:52.254744 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:52 crc kubenswrapper[4778]: E0312 13:11:52.254905 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.266311 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.268418 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.270738 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.280747 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.292675 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.307581 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.324467 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.357785 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: E0312 13:11:52.369865 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.372557 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.393029 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.405457 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.414587 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.427717 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.439940 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.452766 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.469678 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.484869 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.504154 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:49Z\\\",\\\"message\\\":\\\" network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:49.218951 7264 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f32857b5-f652-4313-a0d7-455c3156dd99\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:52 crc kubenswrapper[4778]: I0312 13:11:52.515914 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:52Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.512108 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.512191 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.512227 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.512244 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.512258 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:53Z","lastTransitionTime":"2026-03-12T13:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:53 crc kubenswrapper[4778]: E0312 13:11:53.528684 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:53Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.533651 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.533706 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.533720 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.533739 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.533751 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:53Z","lastTransitionTime":"2026-03-12T13:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:53 crc kubenswrapper[4778]: E0312 13:11:53.559658 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:53Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.564351 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.564440 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.564452 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.564469 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.564481 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:53Z","lastTransitionTime":"2026-03-12T13:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:53 crc kubenswrapper[4778]: E0312 13:11:53.577795 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:53Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.582825 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.582883 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.582893 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.582909 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.582920 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:53Z","lastTransitionTime":"2026-03-12T13:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:53 crc kubenswrapper[4778]: E0312 13:11:53.596233 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:53Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.600456 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.600502 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.600513 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.600534 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:11:53 crc kubenswrapper[4778]: I0312 13:11:53.600548 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:11:53Z","lastTransitionTime":"2026-03-12T13:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:11:53 crc kubenswrapper[4778]: E0312 13:11:53.613449 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:53Z is after 2025-08-24T17:21:41Z" Mar 12 13:11:53 crc kubenswrapper[4778]: E0312 13:11:53.613611 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:11:54 crc kubenswrapper[4778]: I0312 13:11:54.253754 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:54 crc kubenswrapper[4778]: E0312 13:11:54.253974 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:54 crc kubenswrapper[4778]: I0312 13:11:54.254333 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:54 crc kubenswrapper[4778]: E0312 13:11:54.254414 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:54 crc kubenswrapper[4778]: I0312 13:11:54.254604 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:54 crc kubenswrapper[4778]: E0312 13:11:54.254719 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:54 crc kubenswrapper[4778]: I0312 13:11:54.254598 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:54 crc kubenswrapper[4778]: E0312 13:11:54.254850 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:56 crc kubenswrapper[4778]: I0312 13:11:56.253721 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:56 crc kubenswrapper[4778]: I0312 13:11:56.253726 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:56 crc kubenswrapper[4778]: I0312 13:11:56.253866 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:56 crc kubenswrapper[4778]: E0312 13:11:56.254097 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:11:56 crc kubenswrapper[4778]: E0312 13:11:56.254132 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:56 crc kubenswrapper[4778]: E0312 13:11:56.253863 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:56 crc kubenswrapper[4778]: I0312 13:11:56.254928 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:56 crc kubenswrapper[4778]: E0312 13:11:56.255169 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:57 crc kubenswrapper[4778]: E0312 13:11:57.376984 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:11:58 crc kubenswrapper[4778]: I0312 13:11:58.253153 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:11:58 crc kubenswrapper[4778]: I0312 13:11:58.253252 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:11:58 crc kubenswrapper[4778]: I0312 13:11:58.253342 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:11:58 crc kubenswrapper[4778]: E0312 13:11:58.253493 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:11:58 crc kubenswrapper[4778]: I0312 13:11:58.253567 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:11:58 crc kubenswrapper[4778]: E0312 13:11:58.253687 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:11:58 crc kubenswrapper[4778]: E0312 13:11:58.253839 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:11:58 crc kubenswrapper[4778]: E0312 13:11:58.253965 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:00 crc kubenswrapper[4778]: I0312 13:12:00.253799 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:00 crc kubenswrapper[4778]: I0312 13:12:00.253895 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:00 crc kubenswrapper[4778]: E0312 13:12:00.253936 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:00 crc kubenswrapper[4778]: I0312 13:12:00.253970 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:00 crc kubenswrapper[4778]: E0312 13:12:00.254057 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:00 crc kubenswrapper[4778]: I0312 13:12:00.254156 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:00 crc kubenswrapper[4778]: E0312 13:12:00.254345 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:00 crc kubenswrapper[4778]: E0312 13:12:00.254455 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.164061 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.164314 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164376 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:13:06.16433742 +0000 UTC m=+204.613032826 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164442 4778 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.164447 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164551 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:13:06.164521515 +0000 UTC m=+204.613216931 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164705 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164725 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164739 4778 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164798 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 13:13:06.164788242 +0000 UTC m=+204.613483648 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.164807 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.164870 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.164972 4778 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.165034 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 13:13:06.165022479 +0000 UTC m=+204.613717895 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.165125 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.165152 4778 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.165171 4778 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.165277 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 13:13:06.165253075 +0000 UTC m=+204.613948511 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.252922 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.252955 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.253276 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.253250 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.253322 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.253624 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.253688 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.253818 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.266718 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.266919 4778 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.267026 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs podName:0b59b25a-3acc-4d06-b91d-575f45463520 nodeName:}" failed. No retries permitted until 2026-03-12 13:13:06.266996119 +0000 UTC m=+204.715691555 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs") pod "network-metrics-daemon-rz9vw" (UID: "0b59b25a-3acc-4d06-b91d-575f45463520") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.275861 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:43Z\\\",\\\"message\\\":\\\"file observer\\\\nW0312 13:10:42.840582 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 13:10:42.841010 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 13:10:42.843036 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-461564172/tls.crt::/tmp/serving-cert-461564172/tls.key\\\\\\\"\\\\nI0312 13:10:43.350873 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 13:10:43.364662 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 13:10:43.364721 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 13:10:43.365498 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 13:10:43.365555 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 13:10:43.376143 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0312 13:10:43.376224 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376255 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 13:10:43.376279 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 13:10:43.376301 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 13:10:43.376324 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 13:10:43.376350 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0312 13:10:43.376614 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0312 13:10:43.379532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.303700 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:49Z\\\",\\\"message\\\":\\\" network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:11:49Z is after 2025-08-24T17:21:41Z]\\\\nI0312 13:11:49.218951 7264 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f32857b5-f652-4313-a0d7-455c3156dd99\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-schvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8bcc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.319722 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qdxm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7050ddd9-aa01-4af7-9046-208f85f50a86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af31ab4c27bb06d5a44a1c279e04f1b6f243054e271214ef771db4f0dc65e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jspwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qdxm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.341338 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rsshp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1f8eaa-ac07-4478-be5d-0742de6b43c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f732882ddde9d0d0c1d1ef218276d4e14df3a1b36e4e956912efef4873092b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2a965ab43fc04beab1d08a4b626d0e54db69963e6ca5c498f502f4df90a8057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deca991c71450137550fdb82a01b81aaa63e6be64a6d7a96438f6b3d83a8bb5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ae3ee2ab1f6fdf579b28f6ddf2010f9ac048dec5c7668dc467152185d4e1028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b22a0b8a6e5c59e8195280cbe1579af847c709f8b6245df5a16df5af602f11d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://253b2ec5086a2db48bb42ae6024bab9ca832325f9d96cd6ff6944ded362161e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb8f9537926237c4932ef2a9fb701804e03e132f2f56dd9d0e928b7340b1eeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:11:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4hrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rsshp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.356425 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4dfhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfefcab6-a931-413e-8763-0f63f17911cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eecca419cc264b25f1883aac864cc545f0daf973e3b288bc8ea00a8b91e1f124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssbrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4dfhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.374820 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7db5517e-3b54-4509-a2e4-fd8fd83f3b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7b2a12646c299c75286fc95cf2a8fa35fd83031ce3daebec42030d966274ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cf827947c686099ca3c6afad51d866f4ee1d557bc64cc1c70f6213fd4198df2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T13:10:15Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 13:09:44.726810 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 13:09:44.729454 1 observer_polling.go:159] Starting file observer\\\\nI0312 13:09:44.783273 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 13:09:44.796968 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0312 13:10:15.116783 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:10:14Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d772ee1ff9d986b4311494a08c8763bd91704fda6cd9c6f067c98205a4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8141466f1a3447b31eeaeb92f1b2ac9e8ddef4ba3e9a60f2ce9a775c3cce0a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5918a46253f4d68b9bc62ba4357dd2ae6baff245e6b4ca06e44eb7e9b7af9df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: E0312 13:12:02.378018 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.394049 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.411244 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fhcz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T13:11:46Z\\\",\\\"message\\\":\\\"2026-03-12T13:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972\\\\n2026-03-12T13:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_415a65bb-5a20-4f1d-953d-70a2be5bd972 to /host/opt/cni/bin/\\\\n2026-03-12T13:11:01Z [verbose] multus-daemon started\\\\n2026-03-12T13:11:01Z [verbose] Readiness Indicator file check\\\\n2026-03-12T13:11:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-762lp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fhcz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.428178 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de004a2f-3061-4aae-aa57-389219c71023\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478fb92ee4748af677ac761928a4173b506a3e56cf622279e2b2a0e322d4aef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d377b0d5d0a854761257d7bc21a111aed96f85d302bf0c024e021f04cc555fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g92p9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sww7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.441055 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24438fc6-dab0-4a9e-8b97-2532da76d9cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19a882eebff25a2613c68847fcf737648da24f5c8d7648edebb2cb00b6b8950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhn9v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qx88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.457497 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f179525-0f37-4bc0-b853-cdc965ca7af8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfcd0839d0f910ecfd92ecc2db64e4ef06fd90bfda52f24a751f8bf1cf112d8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2656063b4947b28fa0ac1759e349c80fc039346869b1c1d6daad75e93ad407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://981038910a82f8dc9ffff22e601a748571a56541b59c187d9ce4f5d500febd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf73fb2fb0de0ce76c16b7db59c94484062b1f4fc5b6df9633c4740f5bbbc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cf73fb2fb0de0ce76c16b7db59c94484062b1f4fc5b6df9633c4740f5bbbc0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.469721 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cae155c-6ba6-44c1-9814-759fda7c3c86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7698145a8f9a3b12ca021d55f406bc6adf7e139c7e32156ced11a20de194608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ddeb961084ae4041feb2ac05c9fdd2f5c11b4bdc5f5f33878c9ad9e83a2e1a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.488712 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2141104-4933-46fd-9968-0d9498779462\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e77ceb524173a1cdbf6c93b730412dcd8b6aedcee06c40fb757cc8e738e380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b52b689d66d254a521c980330e792ecbcce1102f39f97d6149bf48ad24c5de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc82a592c28b74aef165a164cc4fe4e2e38b6fb48e59f499476a252197e3fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f046d558bf242476327e1ee39ea82ebe104caa081df71caa51a716490d8a6b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e2266711bd32e96e742549772474d9fa43d8f368021e8a7aba3fd1c7b0b87b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a347cade99b7bdbe676a020faf0a90b281672f16c4f580455856786ed781d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3701d4b9c229934646d070a25b4bf944ac544d227ff9ba89fb1885cecfb562de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e0d7207d43b4b2bb79583cb1bb2f31034392eb4193b9b3b2f547f474d335250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T13:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T13:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:09:42Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.499559 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.510761 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b59b25a-3acc-4d06-b91d-575f45463520\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdj5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T13:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rz9vw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.544638 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4059dae21c8267dcec17364a3073a0f25addb6c308620992e9e609b5f5a32e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7ffc17b778f7bd099f0cc70b4e8bcfd77f9d45a9a47de9fedbe270a49f2826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.560864 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:11:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa219bcd71a6f1ae8a889a0409c2bbf61d1efac6a57ad8a22fefe6915e9d15be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.573460 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:02 crc kubenswrapper[4778]: I0312 13:12:02.593228 4778 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T13:10:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb7a47e7099405d73886322b00b013bedee4fb573fa60c9b92d6be3311e65c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T13:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:02Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.254558 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.254766 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.816433 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.816475 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.816484 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.816499 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.816512 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:12:03Z","lastTransitionTime":"2026-03-12T13:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.833721 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.839138 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.839237 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.839291 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.839320 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.839356 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:12:03Z","lastTransitionTime":"2026-03-12T13:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.858719 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.863369 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.863449 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.863470 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.863498 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.863513 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:12:03Z","lastTransitionTime":"2026-03-12T13:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.880889 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.886650 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.886710 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.886733 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.886757 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.886774 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:12:03Z","lastTransitionTime":"2026-03-12T13:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.906237 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.910932 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.910980 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.910994 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.911017 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:12:03 crc kubenswrapper[4778]: I0312 13:12:03.911035 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:12:03Z","lastTransitionTime":"2026-03-12T13:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.927936 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T13:12:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9825271f-f529-4477-b3b1-2a00dbf9b03e\\\",\\\"systemUUID\\\":\\\"65870ff3-f0f2-4ca4-b489-075d672e37ad\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T13:12:03Z is after 2025-08-24T17:21:41Z" Mar 12 13:12:03 crc kubenswrapper[4778]: E0312 13:12:03.928108 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:12:04 crc kubenswrapper[4778]: I0312 13:12:04.253483 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:04 crc kubenswrapper[4778]: I0312 13:12:04.253542 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:04 crc kubenswrapper[4778]: I0312 13:12:04.253732 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:04 crc kubenswrapper[4778]: E0312 13:12:04.253732 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:04 crc kubenswrapper[4778]: I0312 13:12:04.253801 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:04 crc kubenswrapper[4778]: E0312 13:12:04.253935 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:04 crc kubenswrapper[4778]: E0312 13:12:04.254085 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:04 crc kubenswrapper[4778]: E0312 13:12:04.254251 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:06 crc kubenswrapper[4778]: I0312 13:12:06.253395 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:06 crc kubenswrapper[4778]: I0312 13:12:06.253499 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:06 crc kubenswrapper[4778]: I0312 13:12:06.253406 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:06 crc kubenswrapper[4778]: I0312 13:12:06.253401 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:06 crc kubenswrapper[4778]: E0312 13:12:06.253669 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:06 crc kubenswrapper[4778]: E0312 13:12:06.253813 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:06 crc kubenswrapper[4778]: E0312 13:12:06.254079 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:06 crc kubenswrapper[4778]: E0312 13:12:06.254145 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:07 crc kubenswrapper[4778]: E0312 13:12:07.379608 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:08 crc kubenswrapper[4778]: I0312 13:12:08.252771 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:08 crc kubenswrapper[4778]: I0312 13:12:08.252850 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:08 crc kubenswrapper[4778]: E0312 13:12:08.252933 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:08 crc kubenswrapper[4778]: I0312 13:12:08.252951 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:08 crc kubenswrapper[4778]: I0312 13:12:08.253016 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:08 crc kubenswrapper[4778]: E0312 13:12:08.253058 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:08 crc kubenswrapper[4778]: E0312 13:12:08.253145 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:08 crc kubenswrapper[4778]: E0312 13:12:08.253352 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:09 crc kubenswrapper[4778]: I0312 13:12:09.270781 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:09 crc kubenswrapper[4778]: I0312 13:12:09.270750 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:09 crc kubenswrapper[4778]: E0312 13:12:09.271074 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:09 crc kubenswrapper[4778]: E0312 13:12:09.271200 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:10 crc kubenswrapper[4778]: I0312 13:12:10.253491 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:10 crc kubenswrapper[4778]: E0312 13:12:10.253627 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:10 crc kubenswrapper[4778]: I0312 13:12:10.253760 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:10 crc kubenswrapper[4778]: E0312 13:12:10.254236 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:11 crc kubenswrapper[4778]: I0312 13:12:11.253711 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:11 crc kubenswrapper[4778]: I0312 13:12:11.253776 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:11 crc kubenswrapper[4778]: E0312 13:12:11.254880 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:11 crc kubenswrapper[4778]: E0312 13:12:11.254933 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.253680 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:12 crc kubenswrapper[4778]: E0312 13:12:12.253883 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.254442 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:12 crc kubenswrapper[4778]: E0312 13:12:12.254652 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.337378 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.33735782 podStartE2EDuration="1m12.33735782s" podCreationTimestamp="2026-03-12 13:11:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.293634319 +0000 UTC m=+150.742329785" watchObservedRunningTime="2026-03-12 13:12:12.33735782 +0000 UTC m=+150.786053216" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.365908 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qdxm2" podStartSLOduration=101.365877553 podStartE2EDuration="1m41.365877553s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.350262843 +0000 UTC m=+150.798958239" watchObservedRunningTime="2026-03-12 13:12:12.365877553 +0000 UTC m=+150.814572949" Mar 12 13:12:12 crc kubenswrapper[4778]: E0312 13:12:12.380118 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.389524 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=20.389507858 podStartE2EDuration="20.389507858s" podCreationTimestamp="2026-03-12 13:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.387894773 +0000 UTC m=+150.836590189" watchObservedRunningTime="2026-03-12 13:12:12.389507858 +0000 UTC m=+150.838203254" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.389811 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rsshp" podStartSLOduration=101.389805836 podStartE2EDuration="1m41.389805836s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.36649883 +0000 UTC m=+150.815194226" watchObservedRunningTime="2026-03-12 13:12:12.389805836 +0000 UTC m=+150.838501232" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.433081 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-fhcz6" podStartSLOduration=101.433058174 podStartE2EDuration="1m41.433058174s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.42155769 +0000 UTC m=+150.870253096" watchObservedRunningTime="2026-03-12 13:12:12.433058174 +0000 UTC m=+150.881753570" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.433775 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sww7j" podStartSLOduration=101.433769354 podStartE2EDuration="1m41.433769354s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.433060524 +0000 UTC m=+150.881755920" watchObservedRunningTime="2026-03-12 13:12:12.433769354 +0000 UTC m=+150.882464750" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.461853 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4dfhs" podStartSLOduration=101.461835304 podStartE2EDuration="1m41.461835304s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.4450112 +0000 UTC m=+150.893706606" watchObservedRunningTime="2026-03-12 13:12:12.461835304 +0000 UTC m=+150.910530700" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.482061 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=20.482030203 podStartE2EDuration="20.482030203s" podCreationTimestamp="2026-03-12 13:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.470276032 +0000 UTC m=+150.918971448" watchObservedRunningTime="2026-03-12 13:12:12.482030203 +0000 UTC m=+150.930725619" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.482390 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=68.482383702 podStartE2EDuration="1m8.482383702s" podCreationTimestamp="2026-03-12 13:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.481515198 +0000 UTC m=+150.930210624" watchObservedRunningTime="2026-03-12 13:12:12.482383702 +0000 UTC m=+150.931079098" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.507551 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=48.5075301 podStartE2EDuration="48.5075301s" podCreationTimestamp="2026-03-12 13:11:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.506285555 +0000 UTC m=+150.954980981" watchObservedRunningTime="2026-03-12 13:12:12.5075301 +0000 UTC m=+150.956225496" Mar 12 13:12:12 crc kubenswrapper[4778]: I0312 13:12:12.535001 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podStartSLOduration=101.534978373 podStartE2EDuration="1m41.534978373s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:12.53487349 +0000 UTC m=+150.983568886" watchObservedRunningTime="2026-03-12 13:12:12.534978373 +0000 UTC m=+150.983673779" Mar 12 13:12:13 crc kubenswrapper[4778]: I0312 13:12:13.253059 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:13 crc kubenswrapper[4778]: I0312 13:12:13.253202 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:13 crc kubenswrapper[4778]: E0312 13:12:13.253288 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:13 crc kubenswrapper[4778]: E0312 13:12:13.253419 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.136581 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.136646 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.136664 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.136692 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.136711 4778 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T13:12:14Z","lastTransitionTime":"2026-03-12T13:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.197724 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv"] Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.198557 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.200817 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.202281 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.202959 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.203713 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.239153 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.239281 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.239405 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.239490 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.239659 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.253385 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.253440 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:14 crc kubenswrapper[4778]: E0312 13:12:14.253688 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:14 crc kubenswrapper[4778]: E0312 13:12:14.253919 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.315667 4778 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.337741 4778 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.340879 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.340940 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.341013 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.341039 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.341060 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.341053 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.341256 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.342123 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.348938 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.364165 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g7sbv\" (UID: \"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: I0312 13:12:14.518687 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" Mar 12 13:12:14 crc kubenswrapper[4778]: W0312 13:12:14.542503 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d96bbda_3294_4e24_a2d9_c7dd7eef5d9b.slice/crio-1e50805f4f07646798c1339f11688835c3bb1fe857559cf2e9ac127fdcd33eb5 WatchSource:0}: Error finding container 1e50805f4f07646798c1339f11688835c3bb1fe857559cf2e9ac127fdcd33eb5: Status 404 returned error can't find the container with id 1e50805f4f07646798c1339f11688835c3bb1fe857559cf2e9ac127fdcd33eb5 Mar 12 13:12:15 crc kubenswrapper[4778]: I0312 13:12:15.020678 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" event={"ID":"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b","Type":"ContainerStarted","Data":"9b676c552f719540de88e78b618b86242e3d091cf2c5597d9b0b64928d91e299"} Mar 12 13:12:15 crc kubenswrapper[4778]: I0312 13:12:15.020733 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" event={"ID":"7d96bbda-3294-4e24-a2d9-c7dd7eef5d9b","Type":"ContainerStarted","Data":"1e50805f4f07646798c1339f11688835c3bb1fe857559cf2e9ac127fdcd33eb5"} Mar 12 13:12:15 crc kubenswrapper[4778]: I0312 13:12:15.252756 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:15 crc kubenswrapper[4778]: E0312 13:12:15.252865 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:15 crc kubenswrapper[4778]: I0312 13:12:15.253095 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:15 crc kubenswrapper[4778]: E0312 13:12:15.253703 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:16 crc kubenswrapper[4778]: I0312 13:12:16.253698 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:16 crc kubenswrapper[4778]: I0312 13:12:16.254363 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:12:16 crc kubenswrapper[4778]: E0312 13:12:16.254648 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8bcc9_openshift-ovn-kubernetes(65cd795e-eb6e-4995-a4c1-9dea6f425ac5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" Mar 12 13:12:16 crc kubenswrapper[4778]: E0312 13:12:16.254680 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:16 crc kubenswrapper[4778]: I0312 13:12:16.255289 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:16 crc kubenswrapper[4778]: E0312 13:12:16.255646 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:17 crc kubenswrapper[4778]: I0312 13:12:17.253852 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:17 crc kubenswrapper[4778]: E0312 13:12:17.254173 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:17 crc kubenswrapper[4778]: I0312 13:12:17.254774 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:17 crc kubenswrapper[4778]: E0312 13:12:17.256250 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:17 crc kubenswrapper[4778]: E0312 13:12:17.381777 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:18 crc kubenswrapper[4778]: I0312 13:12:18.252861 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:18 crc kubenswrapper[4778]: I0312 13:12:18.252904 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:18 crc kubenswrapper[4778]: E0312 13:12:18.253105 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:18 crc kubenswrapper[4778]: E0312 13:12:18.253320 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:19 crc kubenswrapper[4778]: I0312 13:12:19.252852 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:19 crc kubenswrapper[4778]: I0312 13:12:19.252887 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:19 crc kubenswrapper[4778]: E0312 13:12:19.253034 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:19 crc kubenswrapper[4778]: E0312 13:12:19.253171 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:20 crc kubenswrapper[4778]: I0312 13:12:20.253382 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:20 crc kubenswrapper[4778]: E0312 13:12:20.253588 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:20 crc kubenswrapper[4778]: I0312 13:12:20.254029 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:20 crc kubenswrapper[4778]: E0312 13:12:20.254247 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:21 crc kubenswrapper[4778]: I0312 13:12:21.252759 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:21 crc kubenswrapper[4778]: E0312 13:12:21.252886 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:21 crc kubenswrapper[4778]: I0312 13:12:21.252978 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:21 crc kubenswrapper[4778]: E0312 13:12:21.253171 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:22 crc kubenswrapper[4778]: I0312 13:12:22.253037 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:22 crc kubenswrapper[4778]: I0312 13:12:22.256325 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:22 crc kubenswrapper[4778]: E0312 13:12:22.256315 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:22 crc kubenswrapper[4778]: E0312 13:12:22.256480 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:22 crc kubenswrapper[4778]: E0312 13:12:22.382207 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:23 crc kubenswrapper[4778]: I0312 13:12:23.253645 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:23 crc kubenswrapper[4778]: I0312 13:12:23.253666 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:23 crc kubenswrapper[4778]: E0312 13:12:23.254528 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:23 crc kubenswrapper[4778]: E0312 13:12:23.254612 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:24 crc kubenswrapper[4778]: I0312 13:12:24.253339 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:24 crc kubenswrapper[4778]: E0312 13:12:24.253471 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:24 crc kubenswrapper[4778]: I0312 13:12:24.253524 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:24 crc kubenswrapper[4778]: E0312 13:12:24.253666 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:25 crc kubenswrapper[4778]: I0312 13:12:25.253845 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:25 crc kubenswrapper[4778]: E0312 13:12:25.253983 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:25 crc kubenswrapper[4778]: I0312 13:12:25.253866 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:25 crc kubenswrapper[4778]: E0312 13:12:25.255022 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:26 crc kubenswrapper[4778]: I0312 13:12:26.253901 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:26 crc kubenswrapper[4778]: I0312 13:12:26.254634 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:26 crc kubenswrapper[4778]: E0312 13:12:26.254813 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:26 crc kubenswrapper[4778]: E0312 13:12:26.255005 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:27 crc kubenswrapper[4778]: I0312 13:12:27.253767 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:27 crc kubenswrapper[4778]: I0312 13:12:27.253805 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:27 crc kubenswrapper[4778]: E0312 13:12:27.255468 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:27 crc kubenswrapper[4778]: E0312 13:12:27.255596 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:27 crc kubenswrapper[4778]: E0312 13:12:27.384026 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:28 crc kubenswrapper[4778]: I0312 13:12:28.253390 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:28 crc kubenswrapper[4778]: I0312 13:12:28.253446 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:28 crc kubenswrapper[4778]: E0312 13:12:28.253614 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:28 crc kubenswrapper[4778]: E0312 13:12:28.253731 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:29 crc kubenswrapper[4778]: I0312 13:12:29.253638 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:29 crc kubenswrapper[4778]: I0312 13:12:29.253702 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:29 crc kubenswrapper[4778]: E0312 13:12:29.253760 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:29 crc kubenswrapper[4778]: E0312 13:12:29.253906 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:30 crc kubenswrapper[4778]: I0312 13:12:30.253581 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:30 crc kubenswrapper[4778]: I0312 13:12:30.253715 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:30 crc kubenswrapper[4778]: E0312 13:12:30.255268 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:30 crc kubenswrapper[4778]: E0312 13:12:30.255393 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:30 crc kubenswrapper[4778]: I0312 13:12:30.255760 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.110020 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/3.log" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.113610 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerStarted","Data":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.114091 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.139441 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7sbv" podStartSLOduration=120.13942483 podStartE2EDuration="2m0.13942483s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:15.050499707 +0000 UTC m=+153.499195163" watchObservedRunningTime="2026-03-12 13:12:31.13942483 +0000 UTC m=+169.588120226" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.253682 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.253762 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:31 crc kubenswrapper[4778]: E0312 13:12:31.253864 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:31 crc kubenswrapper[4778]: E0312 13:12:31.253952 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.336229 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podStartSLOduration=120.336194089 podStartE2EDuration="2m0.336194089s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:31.144558564 +0000 UTC m=+169.593253960" watchObservedRunningTime="2026-03-12 13:12:31.336194089 +0000 UTC m=+169.784889505" Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.336964 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-rz9vw"] Mar 12 13:12:31 crc kubenswrapper[4778]: I0312 13:12:31.337069 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:31 crc kubenswrapper[4778]: E0312 13:12:31.337172 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:32 crc kubenswrapper[4778]: I0312 13:12:32.254590 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:32 crc kubenswrapper[4778]: E0312 13:12:32.256453 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:32 crc kubenswrapper[4778]: E0312 13:12:32.384619 4778 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 13:12:33 crc kubenswrapper[4778]: I0312 13:12:33.253773 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:33 crc kubenswrapper[4778]: I0312 13:12:33.253838 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:33 crc kubenswrapper[4778]: I0312 13:12:33.253788 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:33 crc kubenswrapper[4778]: E0312 13:12:33.253994 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:33 crc kubenswrapper[4778]: E0312 13:12:33.254102 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:33 crc kubenswrapper[4778]: E0312 13:12:33.254231 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:34 crc kubenswrapper[4778]: I0312 13:12:34.253485 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:34 crc kubenswrapper[4778]: E0312 13:12:34.253685 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:35 crc kubenswrapper[4778]: I0312 13:12:35.254050 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:35 crc kubenswrapper[4778]: I0312 13:12:35.254172 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:35 crc kubenswrapper[4778]: I0312 13:12:35.254229 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:35 crc kubenswrapper[4778]: E0312 13:12:35.254303 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:35 crc kubenswrapper[4778]: E0312 13:12:35.254537 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:35 crc kubenswrapper[4778]: E0312 13:12:35.254442 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:36 crc kubenswrapper[4778]: I0312 13:12:36.253329 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:36 crc kubenswrapper[4778]: E0312 13:12:36.253516 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 13:12:37 crc kubenswrapper[4778]: I0312 13:12:37.253581 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:37 crc kubenswrapper[4778]: I0312 13:12:37.253581 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:37 crc kubenswrapper[4778]: E0312 13:12:37.254586 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rz9vw" podUID="0b59b25a-3acc-4d06-b91d-575f45463520" Mar 12 13:12:37 crc kubenswrapper[4778]: I0312 13:12:37.253598 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:37 crc kubenswrapper[4778]: E0312 13:12:37.254688 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 13:12:37 crc kubenswrapper[4778]: E0312 13:12:37.254843 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 13:12:38 crc kubenswrapper[4778]: I0312 13:12:38.253296 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:12:38 crc kubenswrapper[4778]: I0312 13:12:38.255707 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 13:12:38 crc kubenswrapper[4778]: I0312 13:12:38.256036 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.252915 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.252940 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.252940 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.255007 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.255317 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.255537 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 13:12:39 crc kubenswrapper[4778]: I0312 13:12:39.256995 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.495524 4778 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.545457 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.546324 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.553051 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xz42x"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.554173 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pgrb5"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.554719 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555021 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555357 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555434 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555469 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8dkpx"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555611 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555666 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.555903 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.556086 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.571079 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.571581 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.571791 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.572456 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.572657 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.572890 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573059 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573325 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573445 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573491 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kw4v"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573552 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573597 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573708 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573758 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573792 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573892 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.573930 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.574077 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.574131 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.574300 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.574318 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.574465 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.574757 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.580084 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.582891 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.585139 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.590389 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.603104 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.616502 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.616839 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.617291 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.617560 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.617788 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.616890 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.616949 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.618277 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.618435 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.618573 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.618595 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.618847 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.618933 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-242cb"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.619299 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.619903 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.620677 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.621133 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.621283 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.621359 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.621448 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.621721 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.631075 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.631430 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.632074 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.632617 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.635926 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.636098 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.636500 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.636742 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.636998 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.637220 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.640067 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.644940 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.645498 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.645750 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646005 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646144 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646223 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646382 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646404 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646467 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.646999 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mx6kn"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647014 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647145 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647295 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647406 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647658 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647715 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647769 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647872 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.647932 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.648282 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.655932 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pgrb5"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.656971 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.678605 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.680251 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.680388 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.681727 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.682314 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.710070 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.710100 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.710143 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.710262 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.710332 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.710773 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.711160 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.711530 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.711810 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.712161 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.712324 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.713827 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.715096 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.715671 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.715734 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ms5xq"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.716268 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.716768 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.717075 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721069 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721328 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721554 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721657 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721662 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721613 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721822 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.721909 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722055 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722207 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722331 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722429 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722514 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722683 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.722925 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.723031 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.723179 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.729001 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.729575 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.730874 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kw4v"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.731259 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.731398 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.731791 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732019 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732402 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-images\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732528 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/de34cf46-4b6a-4f7a-8225-eb77bec57450-node-pullsecrets\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732625 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t5m5\" (UniqueName: \"kubernetes.io/projected/53a87d9e-095f-4669-b121-0b2c88e5fabb-kube-api-access-2t5m5\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732712 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k56zt\" (UniqueName: \"kubernetes.io/projected/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-kube-api-access-k56zt\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732820 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.733626 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-serving-cert\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.733778 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.733873 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.733961 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-serving-cert\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734041 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734123 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-etcd-serving-ca\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734412 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-client-ca\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734505 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xw2\" (UniqueName: \"kubernetes.io/projected/f56ab022-7fcd-406c-b308-b8d5f93a8b55-kube-api-access-b7xw2\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734593 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734726 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36ec67c-df24-46ce-94b9-10619822c15a-audit-dir\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734797 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrlvl\" (UniqueName: \"kubernetes.io/projected/f36ec67c-df24-46ce-94b9-10619822c15a-kube-api-access-xrlvl\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732546 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-xwwxp"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.732596 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.734973 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735833 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-config\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735854 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-config\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735891 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56ab022-7fcd-406c-b308-b8d5f93a8b55-serving-cert\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735910 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-etcd-client\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735926 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-serving-cert\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735960 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-machine-approver-tls\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735978 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-client-ca\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.735998 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736014 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-config\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736050 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736066 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-encryption-config\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736081 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-audit\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736095 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-service-ca-bundle\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736110 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736128 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njgc8\" (UniqueName: \"kubernetes.io/projected/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-kube-api-access-njgc8\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736150 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9bl9\" (UniqueName: \"kubernetes.io/projected/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-kube-api-access-c9bl9\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736167 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7cpc\" (UniqueName: \"kubernetes.io/projected/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-kube-api-access-p7cpc\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736212 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736233 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736250 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53a87d9e-095f-4669-b121-0b2c88e5fabb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736298 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736322 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-image-import-ca\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736337 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-auth-proxy-config\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736384 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736402 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-config\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736418 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhhjj\" (UniqueName: \"kubernetes.io/projected/de34cf46-4b6a-4f7a-8225-eb77bec57450-kube-api-access-rhhjj\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736454 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736474 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e9b292b2-1928-45d2-ad7f-8d510ebaa771-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tplzm\" (UID: \"e9b292b2-1928-45d2-ad7f-8d510ebaa771\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736493 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-serving-cert\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736528 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f56ab022-7fcd-406c-b308-b8d5f93a8b55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736548 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736565 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/de34cf46-4b6a-4f7a-8225-eb77bec57450-audit-dir\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736570 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736580 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-config\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736619 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smq82\" (UniqueName: \"kubernetes.io/projected/e9b292b2-1928-45d2-ad7f-8d510ebaa771-kube-api-access-smq82\") pod \"cluster-samples-operator-665b6dd947-tplzm\" (UID: \"e9b292b2-1928-45d2-ad7f-8d510ebaa771\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736635 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-audit-policies\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736650 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-config\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736700 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a87d9e-095f-4669-b121-0b2c88e5fabb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.736716 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkkl7\" (UniqueName: \"kubernetes.io/projected/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-kube-api-access-pkkl7\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737014 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ww8lt"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737067 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737282 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737447 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737642 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737763 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.737926 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.738406 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.739631 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2wqm5"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.739966 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.740177 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.740309 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.743949 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.744461 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.745844 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-98lbj"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.746395 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.746712 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.747497 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.748552 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.749933 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vpp8t"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.750311 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.751038 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.752359 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.753763 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xz42x"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.755276 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.755915 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.756441 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.756468 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.756627 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxrx4"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.756985 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.757524 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2z5gg"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.758405 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.758636 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bdcvl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.759368 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.759697 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.760199 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.760752 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.761317 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.761854 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.763438 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qf4nv"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.763565 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.764924 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555352-q7fvr"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.765069 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.765375 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.770631 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.771611 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pg48j"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.772938 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.781917 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.785003 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2wqm5"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.788480 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xwwxp"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.789292 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.790056 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.791111 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8dkpx"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.792105 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ww8lt"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.793139 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.794145 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.795204 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-242cb"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.796224 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.797236 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.798263 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vpp8t"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.799303 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2z5gg"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.800606 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mx6kn"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.801706 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-98lbj"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.803598 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-8zmxq"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.804510 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.804564 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-d562t"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.805235 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.805568 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.806837 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.807862 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8zmxq"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.808864 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.809422 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.809955 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bdcvl"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.811246 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.812135 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.813217 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-d562t"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.814494 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxrx4"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.815804 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pg48j"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.817235 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.818289 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.819507 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.820602 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.822089 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.823444 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.824630 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555352-q7fvr"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.827057 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cp2lw"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.828118 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.828328 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cp2lw"] Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.828463 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838233 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9bl9\" (UniqueName: \"kubernetes.io/projected/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-kube-api-access-c9bl9\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838277 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7cpc\" (UniqueName: \"kubernetes.io/projected/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-kube-api-access-p7cpc\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838306 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98vx9\" (UniqueName: \"kubernetes.io/projected/84bb574a-c91e-4720-83c6-6c47c9344ad2-kube-api-access-98vx9\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838328 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838347 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqbjv\" (UniqueName: \"kubernetes.io/projected/c825022c-79bc-44ae-bc64-ee9614aafe25-kube-api-access-rqbjv\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838372 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838400 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838434 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53a87d9e-095f-4669-b121-0b2c88e5fabb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838462 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/78cc82c7-719e-43ad-926f-a387e0845219-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838503 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838527 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-audit-policies\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838565 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-image-import-ca\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838590 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-auth-proxy-config\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838624 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838648 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-serving-cert\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838675 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-service-ca\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838702 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-config\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838726 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhhjj\" (UniqueName: \"kubernetes.io/projected/de34cf46-4b6a-4f7a-8225-eb77bec57450-kube-api-access-rhhjj\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838747 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838771 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838793 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78cc82c7-719e-43ad-926f-a387e0845219-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838822 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e9b292b2-1928-45d2-ad7f-8d510ebaa771-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tplzm\" (UID: \"e9b292b2-1928-45d2-ad7f-8d510ebaa771\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838852 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-serving-cert\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838878 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f56ab022-7fcd-406c-b308-b8d5f93a8b55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838902 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838928 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-console-config\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838949 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bb574a-c91e-4720-83c6-6c47c9344ad2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838968 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk6h6\" (UniqueName: \"kubernetes.io/projected/5bb00a46-7425-4d14-a10c-779a5036bba6-kube-api-access-rk6h6\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.838985 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78cc82c7-719e-43ad-926f-a387e0845219-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839005 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/de34cf46-4b6a-4f7a-8225-eb77bec57450-audit-dir\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839025 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-stats-auth\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839045 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22194d8c-315e-46b9-a23b-daab9d020ce4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839063 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57a417a-5175-4210-98a0-69e579c22e14-config\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839083 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-config\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839114 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c320d1aa-c376-41f2-ac5a-8432120b68e0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kc7s7\" (UID: \"c320d1aa-c376-41f2-ac5a-8432120b68e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839133 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b9j4\" (UniqueName: \"kubernetes.io/projected/8af48f77-25f7-49ca-8bcb-2481aa72ee66-kube-api-access-2b9j4\") pod \"downloads-7954f5f757-mx6kn\" (UID: \"8af48f77-25f7-49ca-8bcb-2481aa72ee66\") " pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839152 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22194d8c-315e-46b9-a23b-daab9d020ce4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839169 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839215 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smq82\" (UniqueName: \"kubernetes.io/projected/e9b292b2-1928-45d2-ad7f-8d510ebaa771-kube-api-access-smq82\") pod \"cluster-samples-operator-665b6dd947-tplzm\" (UID: \"e9b292b2-1928-45d2-ad7f-8d510ebaa771\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839236 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-audit-policies\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839258 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f57a417a-5175-4210-98a0-69e579c22e14-serving-cert\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839278 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-config\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839296 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839315 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f57a417a-5175-4210-98a0-69e579c22e14-trusted-ca\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839342 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-proxy-tls\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839360 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfpd8\" (UniqueName: \"kubernetes.io/projected/c320d1aa-c376-41f2-ac5a-8432120b68e0-kube-api-access-lfpd8\") pod \"package-server-manager-789f6589d5-kc7s7\" (UID: \"c320d1aa-c376-41f2-ac5a-8432120b68e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839386 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a87d9e-095f-4669-b121-0b2c88e5fabb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839405 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkkl7\" (UniqueName: \"kubernetes.io/projected/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-kube-api-access-pkkl7\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839411 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839423 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx5wn\" (UniqueName: \"kubernetes.io/projected/5c8d947a-b62b-4eb9-81d7-94530285e8dc-kube-api-access-bx5wn\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839472 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/de34cf46-4b6a-4f7a-8225-eb77bec57450-audit-dir\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839511 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84bb574a-c91e-4720-83c6-6c47c9344ad2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839535 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-image-import-ca\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839584 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839614 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-images\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839684 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/de34cf46-4b6a-4f7a-8225-eb77bec57450-node-pullsecrets\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839703 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t5m5\" (UniqueName: \"kubernetes.io/projected/53a87d9e-095f-4669-b121-0b2c88e5fabb-kube-api-access-2t5m5\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.839726 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5gts\" (UniqueName: \"kubernetes.io/projected/78cc82c7-719e-43ad-926f-a387e0845219-kube-api-access-n5gts\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840288 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-auth-proxy-config\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840459 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a87d9e-095f-4669-b121-0b2c88e5fabb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840606 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-config\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840745 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-config\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840762 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840789 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/de34cf46-4b6a-4f7a-8225-eb77bec57450-node-pullsecrets\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840854 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmb4r\" (UniqueName: \"kubernetes.io/projected/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-kube-api-access-pmb4r\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840897 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k56zt\" (UniqueName: \"kubernetes.io/projected/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-kube-api-access-k56zt\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.840952 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841020 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-images\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841096 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-audit-policies\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841352 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-serving-cert\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841441 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841605 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-config\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841654 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f56ab022-7fcd-406c-b308-b8d5f93a8b55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841805 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.841944 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-oauth-config\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842038 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c8d947a-b62b-4eb9-81d7-94530285e8dc-service-ca-bundle\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842143 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842244 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-serving-cert\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842378 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac3e8bc-e165-45d4-8c32-1ccda9769857-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842473 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-metrics-certs\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842553 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842635 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-etcd-client\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842718 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842804 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-etcd-serving-ca\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842877 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-client-ca\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.842959 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xw2\" (UniqueName: \"kubernetes.io/projected/f56ab022-7fcd-406c-b308-b8d5f93a8b55-kube-api-access-b7xw2\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843034 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843145 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9vx5\" (UniqueName: \"kubernetes.io/projected/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-kube-api-access-t9vx5\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843280 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843360 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-srv-cert\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843422 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-etcd-serving-ca\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843510 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22194d8c-315e-46b9-a23b-daab9d020ce4-config\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843594 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsrcs\" (UniqueName: \"kubernetes.io/projected/f57a417a-5175-4210-98a0-69e579c22e14-kube-api-access-hsrcs\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843700 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36ec67c-df24-46ce-94b9-10619822c15a-audit-dir\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843766 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-client-ca\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843778 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrlvl\" (UniqueName: \"kubernetes.io/projected/f36ec67c-df24-46ce-94b9-10619822c15a-kube-api-access-xrlvl\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844145 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844179 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-config\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844231 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-serving-cert\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844255 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-trusted-ca-bundle\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844281 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0d33ee6-3a31-4464-b401-7469bf04d240-config\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844285 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844311 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-config\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844337 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d33ee6-3a31-4464-b401-7469bf04d240-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844368 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56ab022-7fcd-406c-b308-b8d5f93a8b55-serving-cert\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844393 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-images\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844421 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-etcd-client\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844445 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-serving-cert\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844469 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-machine-approver-tls\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844496 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-client-ca\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844523 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-default-certificate\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844553 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844581 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-config\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844607 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-encryption-config\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844634 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cl6l\" (UniqueName: \"kubernetes.io/projected/f799c7e9-1c31-40bc-9ece-06a086683a98-kube-api-access-6cl6l\") pod \"control-plane-machine-set-operator-78cbb6b69f-zkrqr\" (UID: \"f799c7e9-1c31-40bc-9ece-06a086683a98\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844668 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844695 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac3e8bc-e165-45d4-8c32-1ccda9769857-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844722 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-encryption-config\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844749 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ac3e8bc-e165-45d4-8c32-1ccda9769857-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844773 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-oauth-serving-cert\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844807 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-audit\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844832 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f799c7e9-1c31-40bc-9ece-06a086683a98-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zkrqr\" (UID: \"f799c7e9-1c31-40bc-9ece-06a086683a98\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844860 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0d33ee6-3a31-4464-b401-7469bf04d240-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844883 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-service-ca-bundle\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844910 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844936 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njgc8\" (UniqueName: \"kubernetes.io/projected/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-kube-api-access-njgc8\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844962 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5bb00a46-7425-4d14-a10c-779a5036bba6-audit-dir\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.844992 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5rtt\" (UniqueName: \"kubernetes.io/projected/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-kube-api-access-z5rtt\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.843990 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36ec67c-df24-46ce-94b9-10619822c15a-audit-dir\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.845434 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53a87d9e-095f-4669-b121-0b2c88e5fabb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846035 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846288 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-config\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846342 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846419 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846604 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846720 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-config\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.846897 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-service-ca-bundle\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.847394 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-config\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.847658 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/de34cf46-4b6a-4f7a-8225-eb77bec57450-audit\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.847813 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e9b292b2-1928-45d2-ad7f-8d510ebaa771-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tplzm\" (UID: \"e9b292b2-1928-45d2-ad7f-8d510ebaa771\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.848050 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-serving-cert\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.848101 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.848234 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-client-ca\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.848429 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.848617 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56ab022-7fcd-406c-b308-b8d5f93a8b55-serving-cert\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.849396 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.849399 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-serving-cert\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.849807 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.850239 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-encryption-config\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.850600 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-etcd-client\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.851043 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-machine-approver-tls\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.851944 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.854719 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-serving-cert\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.855850 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.863060 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de34cf46-4b6a-4f7a-8225-eb77bec57450-serving-cert\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.869079 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.888996 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.909318 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.929468 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.945581 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmb4r\" (UniqueName: \"kubernetes.io/projected/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-kube-api-access-pmb4r\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.945640 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/486a990d-7a56-4eea-a44d-d05a412718c2-metrics-tls\") pod \"dns-operator-744455d44c-2z5gg\" (UID: \"486a990d-7a56-4eea-a44d-d05a412718c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.945668 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-csi-data-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.945693 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac3e8bc-e165-45d4-8c32-1ccda9769857-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.945716 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-metrics-certs\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.945740 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946682 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946778 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22194d8c-315e-46b9-a23b-daab9d020ce4-config\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946810 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsrcs\" (UniqueName: \"kubernetes.io/projected/f57a417a-5175-4210-98a0-69e579c22e14-kube-api-access-hsrcs\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946848 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwz6k\" (UniqueName: \"kubernetes.io/projected/12abcb2c-895a-46af-9c26-66e358259ce9-kube-api-access-nwz6k\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946889 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-trusted-ca-bundle\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946916 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d33ee6-3a31-4464-b401-7469bf04d240-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946950 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-profile-collector-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.946979 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-images\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947010 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-encryption-config\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947038 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cl6l\" (UniqueName: \"kubernetes.io/projected/f799c7e9-1c31-40bc-9ece-06a086683a98-kube-api-access-6cl6l\") pod \"control-plane-machine-set-operator-78cbb6b69f-zkrqr\" (UID: \"f799c7e9-1c31-40bc-9ece-06a086683a98\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947070 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ac3e8bc-e165-45d4-8c32-1ccda9769857-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947333 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0d33ee6-3a31-4464-b401-7469bf04d240-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947364 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af58c501-1c93-4f7a-bdf9-1255879aea5a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947393 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5bb00a46-7425-4d14-a10c-779a5036bba6-audit-dir\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947428 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqbjv\" (UniqueName: \"kubernetes.io/projected/c825022c-79bc-44ae-bc64-ee9614aafe25-kube-api-access-rqbjv\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947472 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-audit-policies\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947502 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/78cc82c7-719e-43ad-926f-a387e0845219-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947540 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-serving-cert\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947566 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-service-ca\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947605 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947635 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78cc82c7-719e-43ad-926f-a387e0845219-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947662 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bb574a-c91e-4720-83c6-6c47c9344ad2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947687 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk6h6\" (UniqueName: \"kubernetes.io/projected/5bb00a46-7425-4d14-a10c-779a5036bba6-kube-api-access-rk6h6\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947712 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78cc82c7-719e-43ad-926f-a387e0845219-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947739 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57a417a-5175-4210-98a0-69e579c22e14-config\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947771 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6h7h\" (UniqueName: \"kubernetes.io/projected/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01-kube-api-access-b6h7h\") pod \"auto-csr-approver-29555352-q7fvr\" (UID: \"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01\") " pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947808 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22194d8c-315e-46b9-a23b-daab9d020ce4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947832 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f57a417a-5175-4210-98a0-69e579c22e14-serving-cert\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947857 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f57a417a-5175-4210-98a0-69e579c22e14-trusted-ca\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947883 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a240fd7b-5854-4548-a847-e5590111964b-secret-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947924 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-proxy-tls\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947950 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfpd8\" (UniqueName: \"kubernetes.io/projected/c320d1aa-c376-41f2-ac5a-8432120b68e0-kube-api-access-lfpd8\") pod \"package-server-manager-789f6589d5-kc7s7\" (UID: \"c320d1aa-c376-41f2-ac5a-8432120b68e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.947978 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88x7r\" (UniqueName: \"kubernetes.io/projected/a240fd7b-5854-4548-a847-e5590111964b-kube-api-access-88x7r\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948017 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5gts\" (UniqueName: \"kubernetes.io/projected/78cc82c7-719e-43ad-926f-a387e0845219-kube-api-access-n5gts\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948042 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22194d8c-315e-46b9-a23b-daab9d020ce4-config\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948075 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlpsj\" (UniqueName: \"kubernetes.io/projected/30ff941c-3c4b-4229-af5a-78bb244a385b-kube-api-access-hlpsj\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948126 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9grsg\" (UniqueName: \"kubernetes.io/projected/f1f25dae-f3e4-481d-8451-4851b60b2ec4-kube-api-access-9grsg\") pod \"multus-admission-controller-857f4d67dd-98lbj\" (UID: \"f1f25dae-f3e4-481d-8451-4851b60b2ec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948561 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c8d947a-b62b-4eb9-81d7-94530285e8dc-service-ca-bundle\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948570 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5bb00a46-7425-4d14-a10c-779a5036bba6-audit-dir\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948796 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948831 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-oauth-config\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948859 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-etcd-client\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948878 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9vx5\" (UniqueName: \"kubernetes.io/projected/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-kube-api-access-t9vx5\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948900 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-srv-cert\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948920 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-plugins-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948960 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-serving-cert\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948978 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0d33ee6-3a31-4464-b401-7469bf04d240-config\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.948996 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9k6q\" (UniqueName: \"kubernetes.io/projected/32bf6158-393f-4423-9255-345581ec5bf1-kube-api-access-p9k6q\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949020 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af58c501-1c93-4f7a-bdf9-1255879aea5a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949042 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-default-certificate\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949063 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-registration-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949086 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac3e8bc-e165-45d4-8c32-1ccda9769857-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949102 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-srv-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949119 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfg2m\" (UniqueName: \"kubernetes.io/projected/486a990d-7a56-4eea-a44d-d05a412718c2-kube-api-access-nfg2m\") pod \"dns-operator-744455d44c-2z5gg\" (UID: \"486a990d-7a56-4eea-a44d-d05a412718c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949141 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-oauth-serving-cert\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949160 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f799c7e9-1c31-40bc-9ece-06a086683a98-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zkrqr\" (UID: \"f799c7e9-1c31-40bc-9ece-06a086683a98\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949180 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32bf6158-393f-4423-9255-345581ec5bf1-serving-cert\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949224 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5rtt\" (UniqueName: \"kubernetes.io/projected/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-kube-api-access-z5rtt\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949243 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/30ff941c-3c4b-4229-af5a-78bb244a385b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949259 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32bf6158-393f-4423-9255-345581ec5bf1-config\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949296 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98vx9\" (UniqueName: \"kubernetes.io/projected/84bb574a-c91e-4720-83c6-6c47c9344ad2-kube-api-access-98vx9\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949315 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/30ff941c-3c4b-4229-af5a-78bb244a385b-proxy-tls\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949333 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6t2k\" (UniqueName: \"kubernetes.io/projected/af58c501-1c93-4f7a-bdf9-1255879aea5a-kube-api-access-n6t2k\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949364 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949404 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58p6m\" (UniqueName: \"kubernetes.io/projected/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-kube-api-access-58p6m\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949432 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949453 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-console-config\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949452 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949472 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-socket-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949491 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-stats-auth\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949506 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22194d8c-315e-46b9-a23b-daab9d020ce4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949526 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c320d1aa-c376-41f2-ac5a-8432120b68e0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kc7s7\" (UID: \"c320d1aa-c376-41f2-ac5a-8432120b68e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949525 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-audit-policies\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949543 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-mountpoint-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949623 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b9j4\" (UniqueName: \"kubernetes.io/projected/8af48f77-25f7-49ca-8bcb-2481aa72ee66-kube-api-access-2b9j4\") pod \"downloads-7954f5f757-mx6kn\" (UID: \"8af48f77-25f7-49ca-8bcb-2481aa72ee66\") " pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949662 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949721 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949758 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f1f25dae-f3e4-481d-8451-4851b60b2ec4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-98lbj\" (UID: \"f1f25dae-f3e4-481d-8451-4851b60b2ec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949807 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dnkc\" (UniqueName: \"kubernetes.io/projected/a875bbd5-0126-4d1c-8b7e-97ac32863981-kube-api-access-4dnkc\") pod \"migrator-59844c95c7-r2r62\" (UID: \"a875bbd5-0126-4d1c-8b7e-97ac32863981\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949844 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84bb574a-c91e-4720-83c6-6c47c9344ad2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949972 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.949898 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx5wn\" (UniqueName: \"kubernetes.io/projected/5c8d947a-b62b-4eb9-81d7-94530285e8dc-kube-api-access-bx5wn\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.950054 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c8d947a-b62b-4eb9-81d7-94530285e8dc-service-ca-bundle\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.950165 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-oauth-serving-cert\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.951230 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5bb00a46-7425-4d14-a10c-779a5036bba6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.951679 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-metrics-certs\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.952307 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0d33ee6-3a31-4464-b401-7469bf04d240-config\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.952468 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-oauth-config\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.952506 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-serving-cert\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.952514 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bb574a-c91e-4720-83c6-6c47c9344ad2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.953017 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-console-config\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.953666 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0d33ee6-3a31-4464-b401-7469bf04d240-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.953976 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-encryption-config\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.954158 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.954846 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-stats-auth\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.955265 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-srv-cert\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.955633 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-serving-cert\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.955949 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5bb00a46-7425-4d14-a10c-779a5036bba6-etcd-client\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.956421 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5c8d947a-b62b-4eb9-81d7-94530285e8dc-default-certificate\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.959748 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84bb574a-c91e-4720-83c6-6c47c9344ad2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.963507 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22194d8c-315e-46b9-a23b-daab9d020ce4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.976387 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.979335 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-trusted-ca-bundle\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.989243 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 12 13:12:44 crc kubenswrapper[4778]: I0312 13:12:44.999208 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-service-ca\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.009237 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.023111 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-proxy-tls\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.029991 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.048895 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.050855 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwz6k\" (UniqueName: \"kubernetes.io/projected/12abcb2c-895a-46af-9c26-66e358259ce9-kube-api-access-nwz6k\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.050935 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-profile-collector-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.050991 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af58c501-1c93-4f7a-bdf9-1255879aea5a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051122 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051262 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6h7h\" (UniqueName: \"kubernetes.io/projected/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01-kube-api-access-b6h7h\") pod \"auto-csr-approver-29555352-q7fvr\" (UID: \"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01\") " pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051351 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a240fd7b-5854-4548-a847-e5590111964b-secret-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051451 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88x7r\" (UniqueName: \"kubernetes.io/projected/a240fd7b-5854-4548-a847-e5590111964b-kube-api-access-88x7r\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051521 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlpsj\" (UniqueName: \"kubernetes.io/projected/30ff941c-3c4b-4229-af5a-78bb244a385b-kube-api-access-hlpsj\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051553 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9grsg\" (UniqueName: \"kubernetes.io/projected/f1f25dae-f3e4-481d-8451-4851b60b2ec4-kube-api-access-9grsg\") pod \"multus-admission-controller-857f4d67dd-98lbj\" (UID: \"f1f25dae-f3e4-481d-8451-4851b60b2ec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051618 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-plugins-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051651 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af58c501-1c93-4f7a-bdf9-1255879aea5a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051677 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9k6q\" (UniqueName: \"kubernetes.io/projected/32bf6158-393f-4423-9255-345581ec5bf1-kube-api-access-p9k6q\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051702 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-registration-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051730 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-srv-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051763 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfg2m\" (UniqueName: \"kubernetes.io/projected/486a990d-7a56-4eea-a44d-d05a412718c2-kube-api-access-nfg2m\") pod \"dns-operator-744455d44c-2z5gg\" (UID: \"486a990d-7a56-4eea-a44d-d05a412718c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051794 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32bf6158-393f-4423-9255-345581ec5bf1-serving-cert\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051826 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/30ff941c-3c4b-4229-af5a-78bb244a385b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051871 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/30ff941c-3c4b-4229-af5a-78bb244a385b-proxy-tls\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051891 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-plugins-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051899 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32bf6158-393f-4423-9255-345581ec5bf1-config\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051959 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-registration-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.051980 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6t2k\" (UniqueName: \"kubernetes.io/projected/af58c501-1c93-4f7a-bdf9-1255879aea5a-kube-api-access-n6t2k\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052050 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58p6m\" (UniqueName: \"kubernetes.io/projected/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-kube-api-access-58p6m\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052075 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-socket-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052124 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-mountpoint-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052199 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f1f25dae-f3e4-481d-8451-4851b60b2ec4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-98lbj\" (UID: \"f1f25dae-f3e4-481d-8451-4851b60b2ec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052226 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dnkc\" (UniqueName: \"kubernetes.io/projected/a875bbd5-0126-4d1c-8b7e-97ac32863981-kube-api-access-4dnkc\") pod \"migrator-59844c95c7-r2r62\" (UID: \"a875bbd5-0126-4d1c-8b7e-97ac32863981\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-mountpoint-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052241 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-socket-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052287 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/486a990d-7a56-4eea-a44d-d05a412718c2-metrics-tls\") pod \"dns-operator-744455d44c-2z5gg\" (UID: \"486a990d-7a56-4eea-a44d-d05a412718c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052309 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-csi-data-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052476 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-csi-data-dir\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.052981 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/30ff941c-3c4b-4229-af5a-78bb244a385b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.055497 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-profile-collector-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.056299 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a240fd7b-5854-4548-a847-e5590111964b-secret-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.069959 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.089207 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.109296 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.122535 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f57a417a-5175-4210-98a0-69e579c22e14-serving-cert\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.129735 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.140096 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57a417a-5175-4210-98a0-69e579c22e14-config\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.160238 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.169269 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.171342 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f57a417a-5175-4210-98a0-69e579c22e14-trusted-ca\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.179105 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-images\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.190138 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.209016 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.223509 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/78cc82c7-719e-43ad-926f-a387e0845219-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.241141 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.249664 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.249726 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78cc82c7-719e-43ad-926f-a387e0845219-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.253530 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f799c7e9-1c31-40bc-9ece-06a086683a98-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zkrqr\" (UID: \"f799c7e9-1c31-40bc-9ece-06a086683a98\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.269370 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.298400 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.302595 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.309572 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.330470 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.335874 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.354631 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.370155 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.378561 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac3e8bc-e165-45d4-8c32-1ccda9769857-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.390497 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.409934 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.417682 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac3e8bc-e165-45d4-8c32-1ccda9769857-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.429663 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.450306 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.469434 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.478771 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c320d1aa-c376-41f2-ac5a-8432120b68e0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kc7s7\" (UID: \"c320d1aa-c376-41f2-ac5a-8432120b68e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.489711 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.496163 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f1f25dae-f3e4-481d-8451-4851b60b2ec4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-98lbj\" (UID: \"f1f25dae-f3e4-481d-8451-4851b60b2ec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.509457 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.529735 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.549254 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.569139 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.588914 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.609743 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.630450 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.650461 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.670175 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.676983 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af58c501-1c93-4f7a-bdf9-1255879aea5a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.688731 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.693927 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af58c501-1c93-4f7a-bdf9-1255879aea5a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.708852 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.729689 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.749824 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.756519 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/30ff941c-3c4b-4229-af5a-78bb244a385b-proxy-tls\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.769986 4778 request.go:700] Waited for 1.010967119s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.772466 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.789638 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.809610 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.829236 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.849785 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.870413 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.889642 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.909620 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.929295 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.937976 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/486a990d-7a56-4eea-a44d-d05a412718c2-metrics-tls\") pod \"dns-operator-744455d44c-2z5gg\" (UID: \"486a990d-7a56-4eea-a44d-d05a412718c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.949093 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.970368 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 12 13:12:45 crc kubenswrapper[4778]: I0312 13:12:45.989256 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.008906 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.029386 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.049683 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052499 4778 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052509 4778 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052569 4778 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052617 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32bf6158-393f-4423-9255-345581ec5bf1-config podName:32bf6158-393f-4423-9255-345581ec5bf1 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:46.552577335 +0000 UTC m=+185.001272771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/32bf6158-393f-4423-9255-345581ec5bf1-config") pod "service-ca-operator-777779d784-pg48j" (UID: "32bf6158-393f-4423-9255-345581ec5bf1") : failed to sync configmap cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052530 4778 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052689 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume podName:a240fd7b-5854-4548-a847-e5590111964b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:46.552658938 +0000 UTC m=+185.001354364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume") pod "collect-profiles-29555340-7tvjm" (UID: "a240fd7b-5854-4548-a847-e5590111964b") : failed to sync configmap cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052753 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32bf6158-393f-4423-9255-345581ec5bf1-serving-cert podName:32bf6158-393f-4423-9255-345581ec5bf1 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:46.552707169 +0000 UTC m=+185.001402705 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/32bf6158-393f-4423-9255-345581ec5bf1-serving-cert") pod "service-ca-operator-777779d784-pg48j" (UID: "32bf6158-393f-4423-9255-345581ec5bf1") : failed to sync secret cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: E0312 13:12:46.052788 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-srv-cert podName:12abcb2c-895a-46af-9c26-66e358259ce9 nodeName:}" failed. No retries permitted until 2026-03-12 13:12:46.552776751 +0000 UTC m=+185.001472187 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-srv-cert") pod "catalog-operator-68c6474976-xcfg6" (UID: "12abcb2c-895a-46af-9c26-66e358259ce9") : failed to sync secret cache: timed out waiting for the condition Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.069042 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.089772 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.109314 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.129273 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.150031 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.169283 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.200849 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.209596 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.230475 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.250161 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.271799 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.289996 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.309570 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.328981 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.349665 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.368768 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.388972 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.409157 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.428930 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.449480 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.468971 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.488895 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.509101 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.529492 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.549452 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.569103 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.586790 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-srv-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.586872 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32bf6158-393f-4423-9255-345581ec5bf1-serving-cert\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.586942 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32bf6158-393f-4423-9255-345581ec5bf1-config\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.587328 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.588857 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.589042 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32bf6158-393f-4423-9255-345581ec5bf1-config\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.589834 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.592299 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12abcb2c-895a-46af-9c26-66e358259ce9-srv-cert\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.593681 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32bf6158-393f-4423-9255-345581ec5bf1-serving-cert\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.610215 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.629068 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.648648 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.675001 4778 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.689459 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.738223 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9bl9\" (UniqueName: \"kubernetes.io/projected/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-kube-api-access-c9bl9\") pod \"controller-manager-879f6c89f-pgrb5\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.759621 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7cpc\" (UniqueName: \"kubernetes.io/projected/f0a4c9a9-348c-4271-b466-4b94f11b2c7c-kube-api-access-p7cpc\") pod \"authentication-operator-69f744f599-8dkpx\" (UID: \"f0a4c9a9-348c-4271-b466-4b94f11b2c7c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.768441 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhhjj\" (UniqueName: \"kubernetes.io/projected/de34cf46-4b6a-4f7a-8225-eb77bec57450-kube-api-access-rhhjj\") pod \"apiserver-76f77b778f-xz42x\" (UID: \"de34cf46-4b6a-4f7a-8225-eb77bec57450\") " pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.787472 4778 request.go:700] Waited for 1.94666237s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.789006 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkkl7\" (UniqueName: \"kubernetes.io/projected/e2967620-e2ce-4763-8a6c-e5a37f3a1f98-kube-api-access-pkkl7\") pod \"machine-api-operator-5694c8668f-242cb\" (UID: \"e2967620-e2ce-4763-8a6c-e5a37f3a1f98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.809470 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t5m5\" (UniqueName: \"kubernetes.io/projected/53a87d9e-095f-4669-b121-0b2c88e5fabb-kube-api-access-2t5m5\") pod \"openshift-apiserver-operator-796bbdcf4f-dh8l6\" (UID: \"53a87d9e-095f-4669-b121-0b2c88e5fabb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.817498 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.828384 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smq82\" (UniqueName: \"kubernetes.io/projected/e9b292b2-1928-45d2-ad7f-8d510ebaa771-kube-api-access-smq82\") pod \"cluster-samples-operator-665b6dd947-tplzm\" (UID: \"e9b292b2-1928-45d2-ad7f-8d510ebaa771\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.848912 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k56zt\" (UniqueName: \"kubernetes.io/projected/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-kube-api-access-k56zt\") pod \"route-controller-manager-6576b87f9c-zpgxh\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.853059 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.860515 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.862861 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xw2\" (UniqueName: \"kubernetes.io/projected/f56ab022-7fcd-406c-b308-b8d5f93a8b55-kube-api-access-b7xw2\") pod \"openshift-config-operator-7777fb866f-x4bxj\" (UID: \"f56ab022-7fcd-406c-b308-b8d5f93a8b55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.871085 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.899730 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrlvl\" (UniqueName: \"kubernetes.io/projected/f36ec67c-df24-46ce-94b9-10619822c15a-kube-api-access-xrlvl\") pod \"oauth-openshift-558db77b4-5kw4v\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.919339 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njgc8\" (UniqueName: \"kubernetes.io/projected/4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd-kube-api-access-njgc8\") pod \"machine-approver-56656f9798-qxqsb\" (UID: \"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.967895 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmb4r\" (UniqueName: \"kubernetes.io/projected/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-kube-api-access-pmb4r\") pod \"marketplace-operator-79b997595-2wqm5\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:46 crc kubenswrapper[4778]: I0312 13:12:46.997572 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.007256 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d33ee6-3a31-4464-b401-7469bf04d240-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dqxml\" (UID: \"a0d33ee6-3a31-4464-b401-7469bf04d240\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.011610 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.022037 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsrcs\" (UniqueName: \"kubernetes.io/projected/f57a417a-5175-4210-98a0-69e579c22e14-kube-api-access-hsrcs\") pod \"console-operator-58897d9998-ww8lt\" (UID: \"f57a417a-5175-4210-98a0-69e579c22e14\") " pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.031079 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cl6l\" (UniqueName: \"kubernetes.io/projected/f799c7e9-1c31-40bc-9ece-06a086683a98-kube-api-access-6cl6l\") pod \"control-plane-machine-set-operator-78cbb6b69f-zkrqr\" (UID: \"f799c7e9-1c31-40bc-9ece-06a086683a98\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.038310 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.046528 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.050651 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqbjv\" (UniqueName: \"kubernetes.io/projected/c825022c-79bc-44ae-bc64-ee9614aafe25-kube-api-access-rqbjv\") pod \"console-f9d7485db-xwwxp\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.065402 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ac3e8bc-e165-45d4-8c32-1ccda9769857-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qtkq6\" (UID: \"0ac3e8bc-e165-45d4-8c32-1ccda9769857\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.079657 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-242cb"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.084805 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5gts\" (UniqueName: \"kubernetes.io/projected/78cc82c7-719e-43ad-926f-a387e0845219-kube-api-access-n5gts\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:47 crc kubenswrapper[4778]: W0312 13:12:47.091731 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2967620_e2ce_4763_8a6c_e5a37f3a1f98.slice/crio-ebee3cbb87a7a15df3d9290e795eb64729bc3bba990be43c53ba64e5f73c2ce6 WatchSource:0}: Error finding container ebee3cbb87a7a15df3d9290e795eb64729bc3bba990be43c53ba64e5f73c2ce6: Status 404 returned error can't find the container with id ebee3cbb87a7a15df3d9290e795eb64729bc3bba990be43c53ba64e5f73c2ce6 Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.103741 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78cc82c7-719e-43ad-926f-a387e0845219-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q677m\" (UID: \"78cc82c7-719e-43ad-926f-a387e0845219\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.122972 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk6h6\" (UniqueName: \"kubernetes.io/projected/5bb00a46-7425-4d14-a10c-779a5036bba6-kube-api-access-rk6h6\") pod \"apiserver-7bbb656c7d-vnndl\" (UID: \"5bb00a46-7425-4d14-a10c-779a5036bba6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.128037 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.134758 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.144779 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfpd8\" (UniqueName: \"kubernetes.io/projected/c320d1aa-c376-41f2-ac5a-8432120b68e0-kube-api-access-lfpd8\") pod \"package-server-manager-789f6589d5-kc7s7\" (UID: \"c320d1aa-c376-41f2-ac5a-8432120b68e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.145112 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.167893 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98vx9\" (UniqueName: \"kubernetes.io/projected/84bb574a-c91e-4720-83c6-6c47c9344ad2-kube-api-access-98vx9\") pod \"openshift-controller-manager-operator-756b6f6bc6-mtlvl\" (UID: \"84bb574a-c91e-4720-83c6-6c47c9344ad2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.179430 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" event={"ID":"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd","Type":"ContainerStarted","Data":"52a4dad447678979841752bb0c254c6980c88948a0ddfa89288cb5ac62331582"} Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.187240 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.187661 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" event={"ID":"e2967620-e2ce-4763-8a6c-e5a37f3a1f98","Type":"ContainerStarted","Data":"ebee3cbb87a7a15df3d9290e795eb64729bc3bba990be43c53ba64e5f73c2ce6"} Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.191853 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b9j4\" (UniqueName: \"kubernetes.io/projected/8af48f77-25f7-49ca-8bcb-2481aa72ee66-kube-api-access-2b9j4\") pod \"downloads-7954f5f757-mx6kn\" (UID: \"8af48f77-25f7-49ca-8bcb-2481aa72ee66\") " pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.208004 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx5wn\" (UniqueName: \"kubernetes.io/projected/5c8d947a-b62b-4eb9-81d7-94530285e8dc-kube-api-access-bx5wn\") pod \"router-default-5444994796-ms5xq\" (UID: \"5c8d947a-b62b-4eb9-81d7-94530285e8dc\") " pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.211458 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.217025 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.227979 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5rtt\" (UniqueName: \"kubernetes.io/projected/3373fbdf-245c-4e98-8bd7-7ad30eb98d76-kube-api-access-z5rtt\") pod \"olm-operator-6b444d44fb-mfjpc\" (UID: \"3373fbdf-245c-4e98-8bd7-7ad30eb98d76\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.239746 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.243929 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2wqm5"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.244576 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9vx5\" (UniqueName: \"kubernetes.io/projected/8f4e3ccc-83e5-40ae-bac2-a5bb1362a531-kube-api-access-t9vx5\") pod \"machine-config-operator-74547568cd-k6dcl\" (UID: \"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.259613 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pgrb5"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.262532 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22194d8c-315e-46b9-a23b-daab9d020ce4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sskj6\" (UID: \"22194d8c-315e-46b9-a23b-daab9d020ce4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:47 crc kubenswrapper[4778]: W0312 13:12:47.274450 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24f4aaf5_c17b_4cd8_9284_6df37f1c2f2d.slice/crio-9ea9ce91a5458d09f7e543bf678a01cfeb2e8462d6860a8c5523bea49359f807 WatchSource:0}: Error finding container 9ea9ce91a5458d09f7e543bf678a01cfeb2e8462d6860a8c5523bea49359f807: Status 404 returned error can't find the container with id 9ea9ce91a5458d09f7e543bf678a01cfeb2e8462d6860a8c5523bea49359f807 Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.280890 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.286483 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.290427 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwz6k\" (UniqueName: \"kubernetes.io/projected/12abcb2c-895a-46af-9c26-66e358259ce9-kube-api-access-nwz6k\") pod \"catalog-operator-68c6474976-xcfg6\" (UID: \"12abcb2c-895a-46af-9c26-66e358259ce9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.292659 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.298875 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.303913 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6h7h\" (UniqueName: \"kubernetes.io/projected/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01-kube-api-access-b6h7h\") pod \"auto-csr-approver-29555352-q7fvr\" (UID: \"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01\") " pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.304271 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.320159 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.326159 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.328606 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88x7r\" (UniqueName: \"kubernetes.io/projected/a240fd7b-5854-4548-a847-e5590111964b-kube-api-access-88x7r\") pod \"collect-profiles-29555340-7tvjm\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.341512 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.343581 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8dkpx"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.347663 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlpsj\" (UniqueName: \"kubernetes.io/projected/30ff941c-3c4b-4229-af5a-78bb244a385b-kube-api-access-hlpsj\") pod \"machine-config-controller-84d6567774-x26ck\" (UID: \"30ff941c-3c4b-4229-af5a-78bb244a385b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.369764 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.372873 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9grsg\" (UniqueName: \"kubernetes.io/projected/f1f25dae-f3e4-481d-8451-4851b60b2ec4-kube-api-access-9grsg\") pod \"multus-admission-controller-857f4d67dd-98lbj\" (UID: \"f1f25dae-f3e4-481d-8451-4851b60b2ec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.390029 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kw4v"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.393569 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9k6q\" (UniqueName: \"kubernetes.io/projected/32bf6158-393f-4423-9255-345581ec5bf1-kube-api-access-p9k6q\") pod \"service-ca-operator-777779d784-pg48j\" (UID: \"32bf6158-393f-4423-9255-345581ec5bf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.395440 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.410470 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfg2m\" (UniqueName: \"kubernetes.io/projected/486a990d-7a56-4eea-a44d-d05a412718c2-kube-api-access-nfg2m\") pod \"dns-operator-744455d44c-2z5gg\" (UID: \"486a990d-7a56-4eea-a44d-d05a412718c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.423378 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.424562 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58p6m\" (UniqueName: \"kubernetes.io/projected/2be5b8df-aaff-4a2b-9b54-78a7e58bc420-kube-api-access-58p6m\") pod \"csi-hostpathplugin-cp2lw\" (UID: \"2be5b8df-aaff-4a2b-9b54-78a7e58bc420\") " pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.445062 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.452092 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6t2k\" (UniqueName: \"kubernetes.io/projected/af58c501-1c93-4f7a-bdf9-1255879aea5a-kube-api-access-n6t2k\") pod \"kube-storage-version-migrator-operator-b67b599dd-wxkb2\" (UID: \"af58c501-1c93-4f7a-bdf9-1255879aea5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.460384 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xz42x"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.460690 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.478218 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dnkc\" (UniqueName: \"kubernetes.io/projected/a875bbd5-0126-4d1c-8b7e-97ac32863981-kube-api-access-4dnkc\") pod \"migrator-59844c95c7-r2r62\" (UID: \"a875bbd5-0126-4d1c-8b7e-97ac32863981\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.478403 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.481550 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.497062 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.498657 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.503714 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.512422 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513490 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqmqc\" (UniqueName: \"kubernetes.io/projected/138bb189-6182-4210-91a7-140f93f36f81-kube-api-access-bqmqc\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513555 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51ee714f-fb23-4420-9e70-1b3134eea18e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513601 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51ee714f-fb23-4420-9e70-1b3134eea18e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513625 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfp8b\" (UniqueName: \"kubernetes.io/projected/5b3e2f12-fdec-46e9-82b4-6777c07281c6-kube-api-access-qfp8b\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513666 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e2f12-fdec-46e9-82b4-6777c07281c6-tmpfs\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513868 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8qmh\" (UniqueName: \"kubernetes.io/projected/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-kube-api-access-h8qmh\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.513956 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-tls\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.514082 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/30697403-66e5-4f68-8e2f-804017bd9d71-signing-key\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.514114 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/658a38d7-a172-432e-a612-6e8cf83f17a2-serving-cert\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.514153 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbz7v\" (UniqueName: \"kubernetes.io/projected/658a38d7-a172-432e-a612-6e8cf83f17a2-kube-api-access-lbz7v\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.514753 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-certs\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.514880 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnjq5\" (UniqueName: \"kubernetes.io/projected/30697403-66e5-4f68-8e2f-804017bd9d71-kube-api-access-qnjq5\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515331 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-trusted-ca\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515377 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/30697403-66e5-4f68-8e2f-804017bd9d71-signing-cabundle\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515422 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-client\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515466 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-node-bootstrap-token\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515496 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515523 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b3e2f12-fdec-46e9-82b4-6777c07281c6-webhook-cert\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515543 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-service-ca\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515567 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbvrg\" (UniqueName: \"kubernetes.io/projected/4d54f13d-85d8-4c95-acef-fcf9f197769a-kube-api-access-qbvrg\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515613 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-config\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515646 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7a887dd-1794-4d66-90a6-299512f32bd1-cert\") pod \"ingress-canary-d562t\" (UID: \"b7a887dd-1794-4d66-90a6-299512f32bd1\") " pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515669 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kb4k\" (UniqueName: \"kubernetes.io/projected/b7a887dd-1794-4d66-90a6-299512f32bd1-kube-api-access-4kb4k\") pod \"ingress-canary-d562t\" (UID: \"b7a887dd-1794-4d66-90a6-299512f32bd1\") " pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515691 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d54f13d-85d8-4c95-acef-fcf9f197769a-trusted-ca\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515748 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtgbd\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-kube-api-access-mtgbd\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515783 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-bound-sa-token\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515803 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138bb189-6182-4210-91a7-140f93f36f81-config-volume\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515836 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/138bb189-6182-4210-91a7-140f93f36f81-metrics-tls\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515852 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-ca\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515917 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-certificates\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515938 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b3e2f12-fdec-46e9-82b4-6777c07281c6-apiservice-cert\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515960 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4d54f13d-85d8-4c95-acef-fcf9f197769a-metrics-tls\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.515979 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4d54f13d-85d8-4c95-acef-fcf9f197769a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.521665 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml"] Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.524423 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.024396438 +0000 UTC m=+186.473091834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.526530 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.535703 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.624577 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.624739 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.124713792 +0000 UTC m=+186.573409188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625116 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-bound-sa-token\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625176 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138bb189-6182-4210-91a7-140f93f36f81-config-volume\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625230 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/138bb189-6182-4210-91a7-140f93f36f81-metrics-tls\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625252 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-ca\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625395 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-certificates\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625414 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b3e2f12-fdec-46e9-82b4-6777c07281c6-apiservice-cert\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625457 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4d54f13d-85d8-4c95-acef-fcf9f197769a-metrics-tls\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625501 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4d54f13d-85d8-4c95-acef-fcf9f197769a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625548 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51ee714f-fb23-4420-9e70-1b3134eea18e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625567 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqmqc\" (UniqueName: \"kubernetes.io/projected/138bb189-6182-4210-91a7-140f93f36f81-kube-api-access-bqmqc\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625595 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51ee714f-fb23-4420-9e70-1b3134eea18e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625613 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfp8b\" (UniqueName: \"kubernetes.io/projected/5b3e2f12-fdec-46e9-82b4-6777c07281c6-kube-api-access-qfp8b\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625638 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e2f12-fdec-46e9-82b4-6777c07281c6-tmpfs\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625752 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8qmh\" (UniqueName: \"kubernetes.io/projected/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-kube-api-access-h8qmh\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625796 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-tls\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625821 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/30697403-66e5-4f68-8e2f-804017bd9d71-signing-key\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625838 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/658a38d7-a172-432e-a612-6e8cf83f17a2-serving-cert\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625899 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbz7v\" (UniqueName: \"kubernetes.io/projected/658a38d7-a172-432e-a612-6e8cf83f17a2-kube-api-access-lbz7v\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625964 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnjq5\" (UniqueName: \"kubernetes.io/projected/30697403-66e5-4f68-8e2f-804017bd9d71-kube-api-access-qnjq5\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.625981 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-certs\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626029 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-trusted-ca\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626046 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/30697403-66e5-4f68-8e2f-804017bd9d71-signing-cabundle\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626076 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-client\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626236 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-node-bootstrap-token\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626304 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626337 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b3e2f12-fdec-46e9-82b4-6777c07281c6-webhook-cert\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626373 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-service-ca\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626400 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbvrg\" (UniqueName: \"kubernetes.io/projected/4d54f13d-85d8-4c95-acef-fcf9f197769a-kube-api-access-qbvrg\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626527 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-config\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626637 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7a887dd-1794-4d66-90a6-299512f32bd1-cert\") pod \"ingress-canary-d562t\" (UID: \"b7a887dd-1794-4d66-90a6-299512f32bd1\") " pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626656 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d54f13d-85d8-4c95-acef-fcf9f197769a-trusted-ca\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626694 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kb4k\" (UniqueName: \"kubernetes.io/projected/b7a887dd-1794-4d66-90a6-299512f32bd1-kube-api-access-4kb4k\") pod \"ingress-canary-d562t\" (UID: \"b7a887dd-1794-4d66-90a6-299512f32bd1\") " pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.626798 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtgbd\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-kube-api-access-mtgbd\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.627848 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.127834419 +0000 UTC m=+186.576529815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.628656 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-config\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.631507 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-service-ca\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.631703 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-ca\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.631741 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-certificates\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.631809 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138bb189-6182-4210-91a7-140f93f36f81-config-volume\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.633227 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-trusted-ca\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.633823 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51ee714f-fb23-4420-9e70-1b3134eea18e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.634533 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.647625 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/30697403-66e5-4f68-8e2f-804017bd9d71-signing-cabundle\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.648455 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/30697403-66e5-4f68-8e2f-804017bd9d71-signing-key\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.656501 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7a887dd-1794-4d66-90a6-299512f32bd1-cert\") pod \"ingress-canary-d562t\" (UID: \"b7a887dd-1794-4d66-90a6-299512f32bd1\") " pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.658949 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.659884 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5b3e2f12-fdec-46e9-82b4-6777c07281c6-tmpfs\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.660018 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b3e2f12-fdec-46e9-82b4-6777c07281c6-apiservice-cert\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.660292 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/658a38d7-a172-432e-a612-6e8cf83f17a2-etcd-client\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.660412 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/138bb189-6182-4210-91a7-140f93f36f81-metrics-tls\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.660417 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/658a38d7-a172-432e-a612-6e8cf83f17a2-serving-cert\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.660889 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-certs\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.660990 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-tls\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.661330 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4d54f13d-85d8-4c95-acef-fcf9f197769a-metrics-tls\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.662322 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d54f13d-85d8-4c95-acef-fcf9f197769a-trusted-ca\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.663474 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b3e2f12-fdec-46e9-82b4-6777c07281c6-webhook-cert\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.664414 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51ee714f-fb23-4420-9e70-1b3134eea18e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.677448 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtgbd\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-kube-api-access-mtgbd\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.678286 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-node-bootstrap-token\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.683473 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.683546 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-bound-sa-token\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.693319 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.704942 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.717837 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8qmh\" (UniqueName: \"kubernetes.io/projected/d6b107a5-befb-4e43-9aa6-6b66ff686bf0-kube-api-access-h8qmh\") pod \"machine-config-server-qf4nv\" (UID: \"d6b107a5-befb-4e43-9aa6-6b66ff686bf0\") " pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.727766 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.727865 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.227843345 +0000 UTC m=+186.676538741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.728011 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.728369 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.228361959 +0000 UTC m=+186.677057355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.728480 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnjq5\" (UniqueName: \"kubernetes.io/projected/30697403-66e5-4f68-8e2f-804017bd9d71-kube-api-access-qnjq5\") pod \"service-ca-9c57cc56f-vpp8t\" (UID: \"30697403-66e5-4f68-8e2f-804017bd9d71\") " pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.748961 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4d54f13d-85d8-4c95-acef-fcf9f197769a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.768161 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbvrg\" (UniqueName: \"kubernetes.io/projected/4d54f13d-85d8-4c95-acef-fcf9f197769a-kube-api-access-qbvrg\") pod \"ingress-operator-5b745b69d9-srhvx\" (UID: \"4d54f13d-85d8-4c95-acef-fcf9f197769a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.769536 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qf4nv" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.783461 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ww8lt"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.783505 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.791164 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbz7v\" (UniqueName: \"kubernetes.io/projected/658a38d7-a172-432e-a612-6e8cf83f17a2-kube-api-access-lbz7v\") pod \"etcd-operator-b45778765-bdcvl\" (UID: \"658a38d7-a172-432e-a612-6e8cf83f17a2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.813887 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqmqc\" (UniqueName: \"kubernetes.io/projected/138bb189-6182-4210-91a7-140f93f36f81-kube-api-access-bqmqc\") pod \"dns-default-8zmxq\" (UID: \"138bb189-6182-4210-91a7-140f93f36f81\") " pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.823682 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xwwxp"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.830179 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.830770 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.3307004 +0000 UTC m=+186.779395796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.842825 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kb4k\" (UniqueName: \"kubernetes.io/projected/b7a887dd-1794-4d66-90a6-299512f32bd1-kube-api-access-4kb4k\") pod \"ingress-canary-d562t\" (UID: \"b7a887dd-1794-4d66-90a6-299512f32bd1\") " pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.912221 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.923168 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfp8b\" (UniqueName: \"kubernetes.io/projected/5b3e2f12-fdec-46e9-82b4-6777c07281c6-kube-api-access-qfp8b\") pod \"packageserver-d55dfcdfc-5m8sg\" (UID: \"5b3e2f12-fdec-46e9-82b4-6777c07281c6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.923621 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.931687 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:47 crc kubenswrapper[4778]: E0312 13:12:47.932079 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.432064312 +0000 UTC m=+186.880759708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.934222 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl"] Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.941974 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:47 crc kubenswrapper[4778]: I0312 13:12:47.950129 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.032781 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.033360 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.533344474 +0000 UTC m=+186.982039870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.036241 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" Mar 12 13:12:48 crc kubenswrapper[4778]: W0312 13:12:48.044846 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ac3e8bc_e165_45d4_8c32_1ccda9769857.slice/crio-5ec74c62b6d6388b4618514bfbb582c3c32d8d7869b55d6601fae9a433a2c7d7 WatchSource:0}: Error finding container 5ec74c62b6d6388b4618514bfbb582c3c32d8d7869b55d6601fae9a433a2c7d7: Status 404 returned error can't find the container with id 5ec74c62b6d6388b4618514bfbb582c3c32d8d7869b55d6601fae9a433a2c7d7 Mar 12 13:12:48 crc kubenswrapper[4778]: W0312 13:12:48.051200 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78cc82c7_719e_43ad_926f_a387e0845219.slice/crio-3923e5748c928aead4163fbcc025b65be3ed9352674f1cc94d77747ffce996b0 WatchSource:0}: Error finding container 3923e5748c928aead4163fbcc025b65be3ed9352674f1cc94d77747ffce996b0: Status 404 returned error can't find the container with id 3923e5748c928aead4163fbcc025b65be3ed9352674f1cc94d77747ffce996b0 Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.054408 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.112886 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.122944 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-d562t" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.137889 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.138267 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.638255017 +0000 UTC m=+187.086950413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.210848 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" event={"ID":"84bb574a-c91e-4720-83c6-6c47c9344ad2","Type":"ContainerStarted","Data":"ea5df04d080d3ebc3170065546468f5918280472695bcdb160c370eb1e447245"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.215910 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" event={"ID":"a0d33ee6-3a31-4464-b401-7469bf04d240","Type":"ContainerStarted","Data":"fb497e12ecda0a19b04837d910aacaba7192beb584f873e810c0be5a336568fe"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.221313 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" event={"ID":"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd","Type":"ContainerStarted","Data":"d666fb76e95e18d9782e85ddfde87b5fa8c5ca2d8db40c92247c06b4ec5f46f1"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.224246 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" event={"ID":"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d","Type":"ContainerStarted","Data":"013c13acbd136a9ae3c6c39b9470a59aa4ab705637939d6af761af9e92e81b9c"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.224271 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" event={"ID":"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d","Type":"ContainerStarted","Data":"9ea9ce91a5458d09f7e543bf678a01cfeb2e8462d6860a8c5523bea49359f807"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.225586 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.228763 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" event={"ID":"f36ec67c-df24-46ce-94b9-10619822c15a","Type":"ContainerStarted","Data":"5f7362fc7516f559081256deebf693613a994486c74f126dfda003689ad66bff"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.230230 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" event={"ID":"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531","Type":"ContainerStarted","Data":"b9c017a33a60d1e8d34966ea3aaa51776f0f2afffd8b7a60693e8b6355a0b61d"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.238634 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.238915 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.73889994 +0000 UTC m=+187.187595326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.240974 4778 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2wqm5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.241029 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" podUID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.244412 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" event={"ID":"f799c7e9-1c31-40bc-9ece-06a086683a98","Type":"ContainerStarted","Data":"8c7ba888781510a0b2db17d5d30a8c1b81e1a4b73eb8d9cf02595412a85f2fa2"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.250909 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwwxp" event={"ID":"c825022c-79bc-44ae-bc64-ee9614aafe25","Type":"ContainerStarted","Data":"6f20116905733a7dbe8802503613a6b31a51c117f53f02f55e4cace656d26f20"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.324104 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" event={"ID":"5bb00a46-7425-4d14-a10c-779a5036bba6","Type":"ContainerStarted","Data":"49e9008c989ae0099a350bef720c95a9b121ede61cd076c12105761c291b6a47"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.324138 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" event={"ID":"53a87d9e-095f-4669-b121-0b2c88e5fabb","Type":"ContainerStarted","Data":"2c3668c7619927e62a21a085b12fd8d10c5aec63a5140a4a4e25a28dcd904a3d"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.324156 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2z5gg"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.324169 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" event={"ID":"f57a417a-5175-4210-98a0-69e579c22e14","Type":"ContainerStarted","Data":"6fc8679ba8e28a5908ef6b03b6e49fa8044b6a863eb8cf84241ac40e31c137e9"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.324178 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ms5xq" event={"ID":"5c8d947a-b62b-4eb9-81d7-94530285e8dc","Type":"ContainerStarted","Data":"5dc51733fb3b2d426e2a91f0e490de5f7451da1e08244b0836e86ddceab88244"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.329391 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.333767 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" event={"ID":"e2967620-e2ce-4763-8a6c-e5a37f3a1f98","Type":"ContainerStarted","Data":"54c08081b13c792d0c2d32cd39505869a93ea17f71e1ad3e05371ec38310e46c"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.333808 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" event={"ID":"e2967620-e2ce-4763-8a6c-e5a37f3a1f98","Type":"ContainerStarted","Data":"edbd6f9ccea31acf5504e98ff7cf249d797a97e8e3fea3b85457c0a95e2146b1"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.336681 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" event={"ID":"f0a4c9a9-348c-4271-b466-4b94f11b2c7c","Type":"ContainerStarted","Data":"06788d28c2ebf93e6930e30ff45162a3683351862e845c812a0ae45bcf63771c"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.336713 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" event={"ID":"f0a4c9a9-348c-4271-b466-4b94f11b2c7c","Type":"ContainerStarted","Data":"5fbdeb3453e8c6fd6686a0aefbf03a12b59237ee3890cbc097329b1e74c72168"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.339477 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.340591 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" event={"ID":"0ac3e8bc-e165-45d4-8c32-1ccda9769857","Type":"ContainerStarted","Data":"5ec74c62b6d6388b4618514bfbb582c3c32d8d7869b55d6601fae9a433a2c7d7"} Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.340761 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.840749457 +0000 UTC m=+187.289444853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.343017 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" event={"ID":"e9b292b2-1928-45d2-ad7f-8d510ebaa771","Type":"ContainerStarted","Data":"342fbb0d2f378f9d2820298cbb5d378552da4f5ad43e0589912c2f9f56bf877e"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.353164 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" event={"ID":"06ffdff1-2f10-4f38-b7fd-b98e883bbc63","Type":"ContainerStarted","Data":"58650ee0315d5aac50c162f5420d39a44557cb90a0d565bd9b299a8e4ee0251d"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.353242 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.358515 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" event={"ID":"78cc82c7-719e-43ad-926f-a387e0845219","Type":"ContainerStarted","Data":"3923e5748c928aead4163fbcc025b65be3ed9352674f1cc94d77747ffce996b0"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.363468 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" event={"ID":"f56ab022-7fcd-406c-b308-b8d5f93a8b55","Type":"ContainerStarted","Data":"3262b29f82e6028fc91389bbd3c254283d8f35d9f9b9a6a3e6427d8f37539038"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.366287 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" event={"ID":"de34cf46-4b6a-4f7a-8225-eb77bec57450","Type":"ContainerStarted","Data":"fa6d0f6d39632346cafc98f07344ee0ccaae7084380a4d10d518d946f29de2d8"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.371942 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" event={"ID":"06bbf7b7-3e40-4aa0-a3db-a56897f5488c","Type":"ContainerStarted","Data":"baecc290d5904f2078cb76008ee3fad41b6baea1393aa1ce14dba9ed727aca24"} Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.372384 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.401355 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mx6kn"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.404450 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.407515 4778 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-zpgxh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.407546 4778 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pgrb5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.407558 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.407588 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.440056 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.442727 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:48.942707907 +0000 UTC m=+187.391403303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.496583 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.511875 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.542059 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.542368 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.042357032 +0000 UTC m=+187.491052428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: W0312 13:12:48.592958 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3373fbdf_245c_4e98_8bd7_7ad30eb98d76.slice/crio-c7dc6a556a0a2c75eb762cfae0c9f40b708a2376fe099f8c49e0c6dd2eec3cb4 WatchSource:0}: Error finding container c7dc6a556a0a2c75eb762cfae0c9f40b708a2376fe099f8c49e0c6dd2eec3cb4: Status 404 returned error can't find the container with id c7dc6a556a0a2c75eb762cfae0c9f40b708a2376fe099f8c49e0c6dd2eec3cb4 Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.643260 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.643679 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.143664684 +0000 UTC m=+187.592360080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.665969 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cp2lw"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.747065 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555352-q7fvr"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.749014 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.749883 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.249871494 +0000 UTC m=+187.698566890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.751341 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.755154 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-98lbj"] Mar 12 13:12:48 crc kubenswrapper[4778]: W0312 13:12:48.776591 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2be5b8df_aaff_4a2b_9b54_78a7e58bc420.slice/crio-29eb382f6373a1343bfe9a9453bc1e1e657ed189d7535c7e86f8a9880039a35d WatchSource:0}: Error finding container 29eb382f6373a1343bfe9a9453bc1e1e657ed189d7535c7e86f8a9880039a35d: Status 404 returned error can't find the container with id 29eb382f6373a1343bfe9a9453bc1e1e657ed189d7535c7e86f8a9880039a35d Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.784064 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.801502 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pg48j"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.805506 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.829531 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.837292 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.853456 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.853780 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.353764479 +0000 UTC m=+187.802459875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.891130 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vpp8t"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.912335 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg"] Mar 12 13:12:48 crc kubenswrapper[4778]: W0312 13:12:48.913744 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32bf6158_393f_4423_9255_345581ec5bf1.slice/crio-87a986e13c20bc428422448d907cce18d39f5ca108a854a847f92de1dc3e4f14 WatchSource:0}: Error finding container 87a986e13c20bc428422448d907cce18d39f5ca108a854a847f92de1dc3e4f14: Status 404 returned error can't find the container with id 87a986e13c20bc428422448d907cce18d39f5ca108a854a847f92de1dc3e4f14 Mar 12 13:12:48 crc kubenswrapper[4778]: W0312 13:12:48.924176 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30697403_66e5_4f68_8e2f_804017bd9d71.slice/crio-86c7d034ba551e67a3168224247007b40f929e6bb1ea121361fae6fadf6d5a7d WatchSource:0}: Error finding container 86c7d034ba551e67a3168224247007b40f929e6bb1ea121361fae6fadf6d5a7d: Status 404 returned error can't find the container with id 86c7d034ba551e67a3168224247007b40f929e6bb1ea121361fae6fadf6d5a7d Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.925967 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bdcvl"] Mar 12 13:12:48 crc kubenswrapper[4778]: I0312 13:12:48.954825 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:48 crc kubenswrapper[4778]: E0312 13:12:48.955133 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.455115022 +0000 UTC m=+187.903810418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: W0312 13:12:49.016503 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf58c501_1c93_4f7a_bdf9_1255879aea5a.slice/crio-2db6fb0d9756a7537c2816ec341008d7b493b3a4c0ad299afe1f4ec5795d8c6c WatchSource:0}: Error finding container 2db6fb0d9756a7537c2816ec341008d7b493b3a4c0ad299afe1f4ec5795d8c6c: Status 404 returned error can't find the container with id 2db6fb0d9756a7537c2816ec341008d7b493b3a4c0ad299afe1f4ec5795d8c6c Mar 12 13:12:49 crc kubenswrapper[4778]: W0312 13:12:49.045076 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod658a38d7_a172_432e_a612_6e8cf83f17a2.slice/crio-b881df9a25cfeae23f99dc528674ad3847aed0b982acf84c8642ced679390821 WatchSource:0}: Error finding container b881df9a25cfeae23f99dc528674ad3847aed0b982acf84c8642ced679390821: Status 404 returned error can't find the container with id b881df9a25cfeae23f99dc528674ad3847aed0b982acf84c8642ced679390821 Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.055838 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.056449 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.556427474 +0000 UTC m=+188.005122870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.056589 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.056931 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.556917737 +0000 UTC m=+188.005613133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: W0312 13:12:49.073571 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b3e2f12_fdec_46e9_82b4_6777c07281c6.slice/crio-2e0ff4520d9ba7a3f7dbf106a17656ddac6fc701b516e47132bdb358130a1ca7 WatchSource:0}: Error finding container 2e0ff4520d9ba7a3f7dbf106a17656ddac6fc701b516e47132bdb358130a1ca7: Status 404 returned error can't find the container with id 2e0ff4520d9ba7a3f7dbf106a17656ddac6fc701b516e47132bdb358130a1ca7 Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.092762 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8zmxq"] Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.140870 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-d562t"] Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.158205 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.158665 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.658645261 +0000 UTC m=+188.107340657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.168095 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx"] Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.261744 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.262617 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.762603648 +0000 UTC m=+188.211299044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.320093 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" podStartSLOduration=138.320073125 podStartE2EDuration="2m18.320073125s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.294854165 +0000 UTC m=+187.743549551" watchObservedRunningTime="2026-03-12 13:12:49.320073125 +0000 UTC m=+187.768768521" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.332733 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-242cb" podStartSLOduration=138.332715861 podStartE2EDuration="2m18.332715861s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.332073343 +0000 UTC m=+187.780768749" watchObservedRunningTime="2026-03-12 13:12:49.332715861 +0000 UTC m=+187.781411257" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.371358 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.371693 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.871663508 +0000 UTC m=+188.320358904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.372314 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.372725 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.872713187 +0000 UTC m=+188.321408583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.374616 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dkpx" podStartSLOduration=138.37459708 podStartE2EDuration="2m18.37459708s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.368681334 +0000 UTC m=+187.817376730" watchObservedRunningTime="2026-03-12 13:12:49.37459708 +0000 UTC m=+187.823292466" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.416859 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" podStartSLOduration=138.41684316 podStartE2EDuration="2m18.41684316s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.411621853 +0000 UTC m=+187.860317269" watchObservedRunningTime="2026-03-12 13:12:49.41684316 +0000 UTC m=+187.865538556" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.425440 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" event={"ID":"f1f25dae-f3e4-481d-8451-4851b60b2ec4","Type":"ContainerStarted","Data":"980e19e30d4bff72a9a681bc0d69f3af04c689f5aa783b2316fc9fd546ecb7fc"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.434125 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" event={"ID":"486a990d-7a56-4eea-a44d-d05a412718c2","Type":"ContainerStarted","Data":"4eb1b68e02020ce18c1ad52f034f48941036714a902575ef015c2beb14437523"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.440303 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" podStartSLOduration=138.440286819 podStartE2EDuration="2m18.440286819s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.439576469 +0000 UTC m=+187.888271875" watchObservedRunningTime="2026-03-12 13:12:49.440286819 +0000 UTC m=+187.888982215" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.448943 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" event={"ID":"06ffdff1-2f10-4f38-b7fd-b98e883bbc63","Type":"ContainerStarted","Data":"c400bf292471252407338cf73137e8439d1cb8b7e278bdf9b5b3d6aae90e459c"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.453454 4778 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-zpgxh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.453536 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.454618 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" event={"ID":"e9b292b2-1928-45d2-ad7f-8d510ebaa771","Type":"ContainerStarted","Data":"aa008fadb4fb70e34dc62c4d4edec425d04c61e4ab43c6809858298b30f7e0b2"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.454652 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" event={"ID":"e9b292b2-1928-45d2-ad7f-8d510ebaa771","Type":"ContainerStarted","Data":"41a87eacc6997d4a29bfede342bec53085fb9489c3d7aec894bc78e028df8993"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.460907 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" event={"ID":"c320d1aa-c376-41f2-ac5a-8432120b68e0","Type":"ContainerStarted","Data":"360be70e7ac69f128077b26106225e7a87b799223ff300c1d31749c0366fb357"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.464997 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" event={"ID":"12abcb2c-895a-46af-9c26-66e358259ce9","Type":"ContainerStarted","Data":"11e6771f10af52fccfb643aa93b8470ecfa9d97140fcc631bc68263159588700"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.469647 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" event={"ID":"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01","Type":"ContainerStarted","Data":"0a2c8918cbacef5d63ed30076a63c59219bb878177978f4909e3ed43cb24db19"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.473118 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.473554 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:49.973535355 +0000 UTC m=+188.422230751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.473714 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwwxp" event={"ID":"c825022c-79bc-44ae-bc64-ee9614aafe25","Type":"ContainerStarted","Data":"4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.482041 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" event={"ID":"a875bbd5-0126-4d1c-8b7e-97ac32863981","Type":"ContainerStarted","Data":"b713fd37291982fd9495c9de930b3c8ac415358b4d18b32e60fd3184fdff6b4b"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.484335 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" podStartSLOduration=138.484320899 podStartE2EDuration="2m18.484320899s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.481814498 +0000 UTC m=+187.930509904" watchObservedRunningTime="2026-03-12 13:12:49.484320899 +0000 UTC m=+187.933016295" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.500792 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" event={"ID":"5b3e2f12-fdec-46e9-82b4-6777c07281c6","Type":"ContainerStarted","Data":"2e0ff4520d9ba7a3f7dbf106a17656ddac6fc701b516e47132bdb358130a1ca7"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.509428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" event={"ID":"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531","Type":"ContainerStarted","Data":"ade109f8b19aa64ebf57e12248a2627abe0ef1352cc1d8d41745864eeb9af532"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.511558 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-d562t" event={"ID":"b7a887dd-1794-4d66-90a6-299512f32bd1","Type":"ContainerStarted","Data":"0e86a2a440437d426c57ac338db66979c28facc8675da461292e9e33f583870a"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.514008 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" event={"ID":"a240fd7b-5854-4548-a847-e5590111964b","Type":"ContainerStarted","Data":"aa7b81ba2e81ec0ae29a489fd430d937b62e78b889977bd807a40c2a99fb3190"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.516987 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mx6kn" event={"ID":"8af48f77-25f7-49ca-8bcb-2481aa72ee66","Type":"ContainerStarted","Data":"c3914cf8da475f980a0549d830bbdc6cf33bc02b55bdcf28abb76023c126912b"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.518100 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" event={"ID":"30ff941c-3c4b-4229-af5a-78bb244a385b","Type":"ContainerStarted","Data":"f438f9315d0d4fa6bf19d831494414bf8f34e034a962108423452f3b7bd9b01d"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.536070 4778 generic.go:334] "Generic (PLEG): container finished" podID="de34cf46-4b6a-4f7a-8225-eb77bec57450" containerID="07d9e33be479df8f83022eb8f6313a4d44e2d9c2c7660f5d3181d691ea1eb84f" exitCode=0 Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.536379 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" event={"ID":"de34cf46-4b6a-4f7a-8225-eb77bec57450","Type":"ContainerDied","Data":"07d9e33be479df8f83022eb8f6313a4d44e2d9c2c7660f5d3181d691ea1eb84f"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.537886 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dh8l6" event={"ID":"53a87d9e-095f-4669-b121-0b2c88e5fabb","Type":"ContainerStarted","Data":"3096dbff8d31b67e1b4031e888fd2bedf9d23f88e4ca5453fd5c49f6afe677f2"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.541673 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" event={"ID":"f57a417a-5175-4210-98a0-69e579c22e14","Type":"ContainerStarted","Data":"fe2bd48f042fddd9bee53465bcf19ec61fc9f0fa0828cce30aec1f9731b43463"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.542599 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.545053 4778 patch_prober.go:28] interesting pod/console-operator-58897d9998-ww8lt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.547360 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8zmxq" event={"ID":"138bb189-6182-4210-91a7-140f93f36f81","Type":"ContainerStarted","Data":"0f2312151b85a47921adb39c7523ee4c5749d688fc423f8acebe62d440387cd3"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.546174 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" podUID="f57a417a-5175-4210-98a0-69e579c22e14" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.554502 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" event={"ID":"f799c7e9-1c31-40bc-9ece-06a086683a98","Type":"ContainerStarted","Data":"922c022f874c83790682af285a756b9800794f02381d448586e308a3a7bcf062"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.564005 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" event={"ID":"3373fbdf-245c-4e98-8bd7-7ad30eb98d76","Type":"ContainerStarted","Data":"c7dc6a556a0a2c75eb762cfae0c9f40b708a2376fe099f8c49e0c6dd2eec3cb4"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.565333 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" event={"ID":"32bf6158-393f-4423-9255-345581ec5bf1","Type":"ContainerStarted","Data":"87a986e13c20bc428422448d907cce18d39f5ca108a854a847f92de1dc3e4f14"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.572663 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" event={"ID":"2be5b8df-aaff-4a2b-9b54-78a7e58bc420","Type":"ContainerStarted","Data":"29eb382f6373a1343bfe9a9453bc1e1e657ed189d7535c7e86f8a9880039a35d"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.574244 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.575233 4778 generic.go:334] "Generic (PLEG): container finished" podID="5bb00a46-7425-4d14-a10c-779a5036bba6" containerID="13cf5abaaeabafcb511ad061e4e3536945084af30fcd8966603d168336a82bfd" exitCode=0 Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.575293 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" event={"ID":"5bb00a46-7425-4d14-a10c-779a5036bba6","Type":"ContainerDied","Data":"13cf5abaaeabafcb511ad061e4e3536945084af30fcd8966603d168336a82bfd"} Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.575172 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.075141856 +0000 UTC m=+188.523837252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.582178 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" event={"ID":"78cc82c7-719e-43ad-926f-a387e0845219","Type":"ContainerStarted","Data":"ecb424cf1c2ce669f160313107801d801e3e3f9d73ad7de65d932577c1112270"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.583366 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" event={"ID":"22194d8c-315e-46b9-a23b-daab9d020ce4","Type":"ContainerStarted","Data":"ab13725b474c3403d18d2ba420cbb9670185cc494c4502b035184e8bb97aaebe"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.584119 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-xwwxp" podStartSLOduration=138.584102078 podStartE2EDuration="2m18.584102078s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.529587194 +0000 UTC m=+187.978282590" watchObservedRunningTime="2026-03-12 13:12:49.584102078 +0000 UTC m=+188.032797474" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.585381 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" event={"ID":"4c56cc09-5f03-4bcc-a4b1-8fed0dcc49bd","Type":"ContainerStarted","Data":"53d4fd937d7c3c94eb542b8870b67caf1f3d97f4ec36b5819e7a8c147b4fd947"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.587917 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" event={"ID":"f36ec67c-df24-46ce-94b9-10619822c15a","Type":"ContainerStarted","Data":"0fe97ea87ef2b2f3106d61689b8bc6549f4b603dd4e79e424ddbe8637587b2f3"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.588161 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.600435 4778 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5kw4v container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" start-of-body= Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.600489 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" podUID="f36ec67c-df24-46ce-94b9-10619822c15a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.621484 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qxqsb" podStartSLOduration=138.62146168 podStartE2EDuration="2m18.62146168s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.620435851 +0000 UTC m=+188.069131257" watchObservedRunningTime="2026-03-12 13:12:49.62146168 +0000 UTC m=+188.070157086" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.622683 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tplzm" podStartSLOduration=138.622674174 podStartE2EDuration="2m18.622674174s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.586510286 +0000 UTC m=+188.035205682" watchObservedRunningTime="2026-03-12 13:12:49.622674174 +0000 UTC m=+188.071369570" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.635323 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ms5xq" event={"ID":"5c8d947a-b62b-4eb9-81d7-94530285e8dc","Type":"ContainerStarted","Data":"873275ea068e9fdf4f4f59d9b6f2f5449326ad5629bee3fa1e6b426dabc39a87"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.655080 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qf4nv" event={"ID":"d6b107a5-befb-4e43-9aa6-6b66ff686bf0","Type":"ContainerStarted","Data":"93b046c6b3a5704968294077b580ebb35334cb91c5606fe736f79233e433b742"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.662005 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zkrqr" podStartSLOduration=138.66198396 podStartE2EDuration="2m18.66198396s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.653102941 +0000 UTC m=+188.101798357" watchObservedRunningTime="2026-03-12 13:12:49.66198396 +0000 UTC m=+188.110679356" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.664659 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" event={"ID":"af58c501-1c93-4f7a-bdf9-1255879aea5a","Type":"ContainerStarted","Data":"2db6fb0d9756a7537c2816ec341008d7b493b3a4c0ad299afe1f4ec5795d8c6c"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.719502 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" event={"ID":"30697403-66e5-4f68-8e2f-804017bd9d71","Type":"ContainerStarted","Data":"86c7d034ba551e67a3168224247007b40f929e6bb1ea121361fae6fadf6d5a7d"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.720365 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.722704 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.222683829 +0000 UTC m=+188.671379225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.741626 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" podStartSLOduration=138.741611962 podStartE2EDuration="2m18.741611962s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.741050366 +0000 UTC m=+188.189745762" watchObservedRunningTime="2026-03-12 13:12:49.741611962 +0000 UTC m=+188.190307358" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.759412 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" event={"ID":"658a38d7-a172-432e-a612-6e8cf83f17a2","Type":"ContainerStarted","Data":"b881df9a25cfeae23f99dc528674ad3847aed0b982acf84c8642ced679390821"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.780546 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" event={"ID":"f56ab022-7fcd-406c-b308-b8d5f93a8b55","Type":"ContainerStarted","Data":"ef89d12de18273d7eb6b5d650d9d0c0d957c7bf2d5d8bd8e1287fde2496cce7e"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.788461 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" event={"ID":"4d54f13d-85d8-4c95-acef-fcf9f197769a","Type":"ContainerStarted","Data":"96c6462d8fa4784945fd07f101d1b63a31611241c1c54492927684396e35d98f"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.796953 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" event={"ID":"06bbf7b7-3e40-4aa0-a3db-a56897f5488c","Type":"ContainerStarted","Data":"1d59a8a49996a8759d79ab136fc846582c037b05e8fb37ea5f281de757e513e2"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.797607 4778 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pgrb5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.797663 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.809224 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" event={"ID":"84bb574a-c91e-4720-83c6-6c47c9344ad2","Type":"ContainerStarted","Data":"dadd3ae69fc3cee25e1016883a7bb995da4c72271918c4c1b226fd8ae2560348"} Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.818111 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.818524 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q677m" podStartSLOduration=138.818494026 podStartE2EDuration="2m18.818494026s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.81756044 +0000 UTC m=+188.266255846" watchObservedRunningTime="2026-03-12 13:12:49.818494026 +0000 UTC m=+188.267189442" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.823112 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.823547 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.323534698 +0000 UTC m=+188.772230094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.897914 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" podStartSLOduration=138.897896581 podStartE2EDuration="2m18.897896581s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.895455712 +0000 UTC m=+188.344151108" watchObservedRunningTime="2026-03-12 13:12:49.897896581 +0000 UTC m=+188.346591977" Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.924136 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:49 crc kubenswrapper[4778]: E0312 13:12:49.926831 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.426805895 +0000 UTC m=+188.875501291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:49 crc kubenswrapper[4778]: I0312 13:12:49.970932 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mtlvl" podStartSLOduration=138.970913417 podStartE2EDuration="2m18.970913417s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.926116506 +0000 UTC m=+188.374811902" watchObservedRunningTime="2026-03-12 13:12:49.970913417 +0000 UTC m=+188.419608803" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.014976 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ms5xq" podStartSLOduration=139.014948446 podStartE2EDuration="2m19.014948446s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:49.972542793 +0000 UTC m=+188.421238189" watchObservedRunningTime="2026-03-12 13:12:50.014948446 +0000 UTC m=+188.463643842" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.027418 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.027696 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.527685595 +0000 UTC m=+188.976380991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.128258 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.128464 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.628440771 +0000 UTC m=+189.077136167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.128911 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.129179 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.629167962 +0000 UTC m=+189.077863348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.231237 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.232584 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.732558582 +0000 UTC m=+189.181253978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.349558 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.350062 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.850046369 +0000 UTC m=+189.298741765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.451466 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.452079 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:50.952058851 +0000 UTC m=+189.400754247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.508345 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.513904 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.513965 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.555400 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.555766 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.055711649 +0000 UTC m=+189.504407035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.656949 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.657630 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.157613747 +0000 UTC m=+189.606309143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.758453 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.758746 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.258736014 +0000 UTC m=+189.707431410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.821085 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37500: no serving certificate available for the kubelet" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.839267 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" event={"ID":"658a38d7-a172-432e-a612-6e8cf83f17a2","Type":"ContainerStarted","Data":"1a94fe45f752a429199972e344bfdc49cc77e2e8aa05a1a5947d14c54d992c55"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.841747 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qf4nv" event={"ID":"d6b107a5-befb-4e43-9aa6-6b66ff686bf0","Type":"ContainerStarted","Data":"62e5d919f802105ea1b3a2039aca2a2150509639e4f2fca11dbcc93d6c21bee0"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.855934 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" event={"ID":"de34cf46-4b6a-4f7a-8225-eb77bec57450","Type":"ContainerStarted","Data":"977bd127f0730bf16814424d6cb4d6c6fccd07e2e253f06e4bf87e29251d2276"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.859770 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" event={"ID":"22194d8c-315e-46b9-a23b-daab9d020ce4","Type":"ContainerStarted","Data":"bc366c6bd3c76384acb924b41a3ec766ea09a641f3808e42f42e2bfc77dd7ebb"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.860657 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.861058 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.361029844 +0000 UTC m=+189.809725320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.863667 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8zmxq" event={"ID":"138bb189-6182-4210-91a7-140f93f36f81","Type":"ContainerStarted","Data":"d44db0e757aeb75338ebe9a677f885286a6821eb8a7260b30e8ada0524b14a29"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.873560 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" event={"ID":"af58c501-1c93-4f7a-bdf9-1255879aea5a","Type":"ContainerStarted","Data":"f2055f809e508fa781aa94aa5bc3f6dc6240d0cc87c21dca861f82b62915db00"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.888654 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" event={"ID":"a240fd7b-5854-4548-a847-e5590111964b","Type":"ContainerStarted","Data":"6e9a4135f2199a3918c9a565e1055b2ed771be6904f7c3aed074108524b55a58"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.894979 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mx6kn" event={"ID":"8af48f77-25f7-49ca-8bcb-2481aa72ee66","Type":"ContainerStarted","Data":"3f895f754eef0bb8583d87125c78a2d88dff6c4008a74e09e3b011b8af8c89ff"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.895290 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.897000 4778 patch_prober.go:28] interesting pod/downloads-7954f5f757-mx6kn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.897268 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mx6kn" podUID="8af48f77-25f7-49ca-8bcb-2481aa72ee66" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.954706 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-bdcvl" podStartSLOduration=139.95468603 podStartE2EDuration="2m19.95468603s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:50.882865908 +0000 UTC m=+189.331561304" watchObservedRunningTime="2026-03-12 13:12:50.95468603 +0000 UTC m=+189.403381426" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.955451 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskj6" podStartSLOduration=139.955443941 podStartE2EDuration="2m19.955443941s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:50.945942404 +0000 UTC m=+189.394637800" watchObservedRunningTime="2026-03-12 13:12:50.955443941 +0000 UTC m=+189.404139337" Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.962420 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:50 crc kubenswrapper[4778]: E0312 13:12:50.964202 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.464173137 +0000 UTC m=+189.912868533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.982321 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" event={"ID":"a875bbd5-0126-4d1c-8b7e-97ac32863981","Type":"ContainerStarted","Data":"2a6f6743e07a93e70f829bc549f2d33e714e88c1eb288f701ddc0ccd95329306"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.982370 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" event={"ID":"a875bbd5-0126-4d1c-8b7e-97ac32863981","Type":"ContainerStarted","Data":"f54de55cfebc9535d07929914957e08a5c6384e4086bd842202052c04c36e88d"} Mar 12 13:12:50 crc kubenswrapper[4778]: I0312 13:12:50.982769 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37502: no serving certificate available for the kubelet" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.023916 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wxkb2" podStartSLOduration=140.023898109 podStartE2EDuration="2m20.023898109s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.022463768 +0000 UTC m=+189.471159164" watchObservedRunningTime="2026-03-12 13:12:51.023898109 +0000 UTC m=+189.472593505" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.043795 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qf4nv" podStartSLOduration=7.043776368 podStartE2EDuration="7.043776368s" podCreationTimestamp="2026-03-12 13:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:50.981963648 +0000 UTC m=+189.430659044" watchObservedRunningTime="2026-03-12 13:12:51.043776368 +0000 UTC m=+189.492471764" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.045837 4778 generic.go:334] "Generic (PLEG): container finished" podID="f56ab022-7fcd-406c-b308-b8d5f93a8b55" containerID="ef89d12de18273d7eb6b5d650d9d0c0d957c7bf2d5d8bd8e1287fde2496cce7e" exitCode=0 Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.046112 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" event={"ID":"f56ab022-7fcd-406c-b308-b8d5f93a8b55","Type":"ContainerDied","Data":"ef89d12de18273d7eb6b5d650d9d0c0d957c7bf2d5d8bd8e1287fde2496cce7e"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.063244 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.075215 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.575179842 +0000 UTC m=+190.023875228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.078606 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.090568 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37510: no serving certificate available for the kubelet" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.090883 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" event={"ID":"30ff941c-3c4b-4229-af5a-78bb244a385b","Type":"ContainerStarted","Data":"736a8018c349cf2341e6f4eb94afbf0f3171b64da1d22d244119a823279f1073"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.090929 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" event={"ID":"30ff941c-3c4b-4229-af5a-78bb244a385b","Type":"ContainerStarted","Data":"51a7a169aacb978114e1d382311c332111172d449603d79426b4fbb9d6a05461"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.106834 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" event={"ID":"c320d1aa-c376-41f2-ac5a-8432120b68e0","Type":"ContainerStarted","Data":"6ac8ee3b3ac245da5d15d00bd6d7d2189b809610a626bbb976c1ed31a09619c0"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.106889 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" event={"ID":"c320d1aa-c376-41f2-ac5a-8432120b68e0","Type":"ContainerStarted","Data":"94fc500339d319a3749865db6827f6e083f34ce60814afcc4595ae3cf7706b0f"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.107282 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.114853 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" podStartSLOduration=140.114825718 podStartE2EDuration="2m20.114825718s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.062615208 +0000 UTC m=+189.511310614" watchObservedRunningTime="2026-03-12 13:12:51.114825718 +0000 UTC m=+189.563521114" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.120827 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2r62" podStartSLOduration=140.120798706 podStartE2EDuration="2m20.120798706s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.102780229 +0000 UTC m=+189.551475625" watchObservedRunningTime="2026-03-12 13:12:51.120798706 +0000 UTC m=+189.569494102" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.128998 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mx6kn" podStartSLOduration=140.128981707 podStartE2EDuration="2m20.128981707s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.127469214 +0000 UTC m=+189.576164610" watchObservedRunningTime="2026-03-12 13:12:51.128981707 +0000 UTC m=+189.577677103" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.141533 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" event={"ID":"f1f25dae-f3e4-481d-8451-4851b60b2ec4","Type":"ContainerStarted","Data":"871c9b8ae46dc5f00029ac0f63b85dcdfbb540de745b55b8fad5197a886b2a5c"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.158244 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" podStartSLOduration=140.158169378 podStartE2EDuration="2m20.158169378s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.15751436 +0000 UTC m=+189.606209756" watchObservedRunningTime="2026-03-12 13:12:51.158169378 +0000 UTC m=+189.606864784" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.169410 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.169976 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.66996306 +0000 UTC m=+190.118658456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.187542 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37516: no serving certificate available for the kubelet" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.211806 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" podStartSLOduration=140.211789548 podStartE2EDuration="2m20.211789548s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.211227552 +0000 UTC m=+189.659922938" watchObservedRunningTime="2026-03-12 13:12:51.211789548 +0000 UTC m=+189.660484944" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.212359 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x26ck" podStartSLOduration=140.212352264 podStartE2EDuration="2m20.212352264s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.188036339 +0000 UTC m=+189.636731735" watchObservedRunningTime="2026-03-12 13:12:51.212352264 +0000 UTC m=+189.661047670" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.230690 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" event={"ID":"5bb00a46-7425-4d14-a10c-779a5036bba6","Type":"ContainerStarted","Data":"f7f91d4ef783aecf99698d30f7a9029765a79e8641baa462e35a9190c654959c"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.231572 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" podStartSLOduration=140.231548194 podStartE2EDuration="2m20.231548194s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.231346528 +0000 UTC m=+189.680041924" watchObservedRunningTime="2026-03-12 13:12:51.231548194 +0000 UTC m=+189.680243590" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.239435 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" event={"ID":"8f4e3ccc-83e5-40ae-bac2-a5bb1362a531","Type":"ContainerStarted","Data":"9abf18e22758bd2509c1b7021d9a9592da359893bd8606322d57431f2b9769fe"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.251976 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" event={"ID":"3373fbdf-245c-4e98-8bd7-7ad30eb98d76","Type":"ContainerStarted","Data":"3463087063cc705e03cb9086a460b8a9bd82be3b339b0c6111d89b673100c8da"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.252591 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.253366 4778 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-mfjpc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.253403 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" podUID="3373fbdf-245c-4e98-8bd7-7ad30eb98d76" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.269642 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-d562t" event={"ID":"b7a887dd-1794-4d66-90a6-299512f32bd1","Type":"ContainerStarted","Data":"c6418e7bd558b405c93379d97cd1366e81f10d02abe7418779ab78424b4846cd"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.275071 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.276391 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.776373116 +0000 UTC m=+190.225068522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.280287 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" podStartSLOduration=140.280259335 podStartE2EDuration="2m20.280259335s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.277214219 +0000 UTC m=+189.725909615" watchObservedRunningTime="2026-03-12 13:12:51.280259335 +0000 UTC m=+189.728954731" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.299095 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" event={"ID":"a0d33ee6-3a31-4464-b401-7469bf04d240","Type":"ContainerStarted","Data":"bc86fb5453be0badafca787854e5384c3f461fa225c7b32e2c3dcc0425f53429"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.305645 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37528: no serving certificate available for the kubelet" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.308026 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-d562t" podStartSLOduration=7.308008446 podStartE2EDuration="7.308008446s" podCreationTimestamp="2026-03-12 13:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.30601232 +0000 UTC m=+189.754707716" watchObservedRunningTime="2026-03-12 13:12:51.308008446 +0000 UTC m=+189.756703842" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.336497 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" event={"ID":"486a990d-7a56-4eea-a44d-d05a412718c2","Type":"ContainerStarted","Data":"14d7b91c05684bc0c7f86d13f791bcbc70c688ba76edaf092234eac29a1a59d0"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.353886 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k6dcl" podStartSLOduration=140.353867527 podStartE2EDuration="2m20.353867527s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.35360511 +0000 UTC m=+189.802300506" watchObservedRunningTime="2026-03-12 13:12:51.353867527 +0000 UTC m=+189.802562923" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.354259 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" podStartSLOduration=140.354253678 podStartE2EDuration="2m20.354253678s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.331267781 +0000 UTC m=+189.779963177" watchObservedRunningTime="2026-03-12 13:12:51.354253678 +0000 UTC m=+189.802949074" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.366231 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" event={"ID":"5b3e2f12-fdec-46e9-82b4-6777c07281c6","Type":"ContainerStarted","Data":"4727c38ef7af2e309805688aef4d4b418cfbd9b0c2cc9738353eb5971d381c94"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.367277 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.370479 4778 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5m8sg container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.370525 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" podUID="5b3e2f12-fdec-46e9-82b4-6777c07281c6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.381297 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.381409 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqxml" podStartSLOduration=140.381389162 podStartE2EDuration="2m20.381389162s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.381081763 +0000 UTC m=+189.829777159" watchObservedRunningTime="2026-03-12 13:12:51.381389162 +0000 UTC m=+189.830084548" Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.386982 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.886953779 +0000 UTC m=+190.335649255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.396231 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" event={"ID":"32bf6158-393f-4423-9255-345581ec5bf1","Type":"ContainerStarted","Data":"0157c9072ec393da7ee9f835843adb2c73eae8a6c1edd543879099d94372fcea"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.404094 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" podStartSLOduration=140.404077701 podStartE2EDuration="2m20.404077701s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.402233389 +0000 UTC m=+189.850928805" watchObservedRunningTime="2026-03-12 13:12:51.404077701 +0000 UTC m=+189.852773097" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.404147 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" event={"ID":"12abcb2c-895a-46af-9c26-66e358259ce9","Type":"ContainerStarted","Data":"e3aab9f9d0a4b329b434850484ecbf71bf33b51d3d3868abd9c9a08033d2cf6c"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.405494 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.413145 4778 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xcfg6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.413210 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" podUID="12abcb2c-895a-46af-9c26-66e358259ce9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.429288 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" event={"ID":"30697403-66e5-4f68-8e2f-804017bd9d71","Type":"ContainerStarted","Data":"d134eddf0f2d2f7ea60b92781b8038b8007282fa3e0456bea6bd85255a9eb032"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.450728 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" event={"ID":"0ac3e8bc-e165-45d4-8c32-1ccda9769857","Type":"ContainerStarted","Data":"f70ecf9dd646270d97ee1c3ccacc37cbf0166dc7139bb6bbe80179622d36c12e"} Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.455597 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37536: no serving certificate available for the kubelet" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.470166 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" podStartSLOduration=140.470144351 podStartE2EDuration="2m20.470144351s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.467586288 +0000 UTC m=+189.916281684" watchObservedRunningTime="2026-03-12 13:12:51.470144351 +0000 UTC m=+189.918839757" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.470619 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pg48j" podStartSLOduration=140.470610593 podStartE2EDuration="2m20.470610593s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.430071062 +0000 UTC m=+189.878766458" watchObservedRunningTime="2026-03-12 13:12:51.470610593 +0000 UTC m=+189.919305989" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.481945 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.482725 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:51.982710823 +0000 UTC m=+190.431406219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.493934 4778 patch_prober.go:28] interesting pod/console-operator-58897d9998-ww8lt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.493986 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" podUID="f57a417a-5175-4210-98a0-69e579c22e14" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.505027 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.505137 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.506548 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.511816 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:51 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:51 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:51 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.511862 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.575996 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qtkq6" podStartSLOduration=140.575977729 podStartE2EDuration="2m20.575977729s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.522797972 +0000 UTC m=+189.971493378" watchObservedRunningTime="2026-03-12 13:12:51.575977729 +0000 UTC m=+190.024673115" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.577447 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vpp8t" podStartSLOduration=140.5774407 podStartE2EDuration="2m20.5774407s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.574712833 +0000 UTC m=+190.023408239" watchObservedRunningTime="2026-03-12 13:12:51.5774407 +0000 UTC m=+190.026136086" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.583721 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.586114 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.086098254 +0000 UTC m=+190.534793750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.615264 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" podStartSLOduration=140.615242964 podStartE2EDuration="2m20.615242964s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.614788201 +0000 UTC m=+190.063483597" watchObservedRunningTime="2026-03-12 13:12:51.615242964 +0000 UTC m=+190.063938370" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.691455 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.691661 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.191640225 +0000 UTC m=+190.640335621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.692306 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.692838 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.192820458 +0000 UTC m=+190.641515854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.701693 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37550: no serving certificate available for the kubelet" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.731380 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" podStartSLOduration=140.731365113 podStartE2EDuration="2m20.731365113s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:51.73055702 +0000 UTC m=+190.179252426" watchObservedRunningTime="2026-03-12 13:12:51.731365113 +0000 UTC m=+190.180060509" Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.794803 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.795298 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.295277772 +0000 UTC m=+190.743973168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.896078 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.896447 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.39643445 +0000 UTC m=+190.845129846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:51 crc kubenswrapper[4778]: I0312 13:12:51.996648 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:51 crc kubenswrapper[4778]: E0312 13:12:51.996975 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.49696092 +0000 UTC m=+190.945656316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.062395 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37566: no serving certificate available for the kubelet" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.098367 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.098681 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.598670163 +0000 UTC m=+191.047365559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.188109 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.188160 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.189119 4778 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-vnndl container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.189156 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" podUID="5bb00a46-7425-4d14-a10c-779a5036bba6" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.199133 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.199490 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.69946677 +0000 UTC m=+191.148162166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.301268 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.301602 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.801587485 +0000 UTC m=+191.250282881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.401961 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.402118 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.902094744 +0000 UTC m=+191.350790140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.402335 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.402632 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:52.902619539 +0000 UTC m=+191.351314925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.502989 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.503372 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.003356865 +0000 UTC m=+191.452052261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.519710 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:52 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:52 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:52 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.519770 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.525437 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" event={"ID":"f56ab022-7fcd-406c-b308-b8d5f93a8b55","Type":"ContainerStarted","Data":"2a2fbf21962ed8b3c22585330718ea835e92053ee374a04976de65692706dcd7"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.554654 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" event={"ID":"de34cf46-4b6a-4f7a-8225-eb77bec57450","Type":"ContainerStarted","Data":"401b5703c41acfddd23d014f272ac8f6b1388f62ecea38307a5473a536d9b0d6"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.574834 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" event={"ID":"4d54f13d-85d8-4c95-acef-fcf9f197769a","Type":"ContainerStarted","Data":"80f2cfadac8d6b03ff7892bf24e4883f713e9acbbedfe471377dd22adfc37de7"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.574887 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-srhvx" event={"ID":"4d54f13d-85d8-4c95-acef-fcf9f197769a","Type":"ContainerStarted","Data":"d466ad97054be4ed37f9fb8b3ec96f64dde9ee13ebda4c3eabdc8fba39283309"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.586171 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8zmxq" event={"ID":"138bb189-6182-4210-91a7-140f93f36f81","Type":"ContainerStarted","Data":"77c33b0ef2ccc7fb4f73af4b872371447a693d9b8d223a80bae6ca75eb6b33b9"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.586446 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-8zmxq" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.590465 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" event={"ID":"2be5b8df-aaff-4a2b-9b54-78a7e58bc420","Type":"ContainerStarted","Data":"f95ce1da1eda4c3816899c45ffc12f805515fb791f219e8cfde75c837042966f"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.596199 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" podStartSLOduration=141.596171628 podStartE2EDuration="2m21.596171628s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:52.595126988 +0000 UTC m=+191.043822394" watchObservedRunningTime="2026-03-12 13:12:52.596171628 +0000 UTC m=+191.044867024" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.604830 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.605309 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.105292134 +0000 UTC m=+191.553987520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.617321 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-98lbj" event={"ID":"f1f25dae-f3e4-481d-8451-4851b60b2ec4","Type":"ContainerStarted","Data":"51cb6200b8285b0484a8c1780537acf8b2d6e7ed966382ba2545a9eb663773f3"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.630066 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-8zmxq" podStartSLOduration=8.630051731 podStartE2EDuration="8.630051731s" podCreationTimestamp="2026-03-12 13:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:52.628610021 +0000 UTC m=+191.077305437" watchObservedRunningTime="2026-03-12 13:12:52.630051731 +0000 UTC m=+191.078747127" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.643150 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2z5gg" event={"ID":"486a990d-7a56-4eea-a44d-d05a412718c2","Type":"ContainerStarted","Data":"4b2262b04363b849751ba177591dd70f0df5ed809dd0d98e5777ad3e10a0be6c"} Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.647179 4778 patch_prober.go:28] interesting pod/downloads-7954f5f757-mx6kn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.647279 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mx6kn" podUID="8af48f77-25f7-49ca-8bcb-2481aa72ee66" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.657780 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mfjpc" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.662121 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-ww8lt" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.668995 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xcfg6" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.705548 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.708655 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.208636333 +0000 UTC m=+191.657331729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.808009 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.811051 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.311036406 +0000 UTC m=+191.759731802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.815645 4778 ???:1] "http: TLS handshake error from 192.168.126.11:37574: no serving certificate available for the kubelet" Mar 12 13:12:52 crc kubenswrapper[4778]: I0312 13:12:52.909192 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:52 crc kubenswrapper[4778]: E0312 13:12:52.909712 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.409691863 +0000 UTC m=+191.858387269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.010413 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.010751 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.510734308 +0000 UTC m=+191.959429704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.110916 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.111458 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.611443763 +0000 UTC m=+192.060139159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.136611 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5m8sg" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.212128 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.212443 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.712430195 +0000 UTC m=+192.161125591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.313675 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.314053 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.814038306 +0000 UTC m=+192.262733702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.374026 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-khr6h"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.374926 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.380076 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.413425 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-khr6h"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.418314 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-catalog-content\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.418350 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-utilities\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.418397 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzdp9\" (UniqueName: \"kubernetes.io/projected/1d185732-cd6b-44c6-b4db-ee9ade00c683-kube-api-access-zzdp9\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.418462 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.418756 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:53.918743653 +0000 UTC m=+192.367439039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.509915 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:53 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:53 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:53 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.509973 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.519530 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.519734 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.019702475 +0000 UTC m=+192.468397871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.519780 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.519831 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-catalog-content\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.519850 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-utilities\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.519889 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzdp9\" (UniqueName: \"kubernetes.io/projected/1d185732-cd6b-44c6-b4db-ee9ade00c683-kube-api-access-zzdp9\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.520212 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.020203559 +0000 UTC m=+192.468898955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.520355 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-catalog-content\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.520407 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-utilities\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.531766 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qx9d8"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.532751 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.542408 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.574031 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzdp9\" (UniqueName: \"kubernetes.io/projected/1d185732-cd6b-44c6-b4db-ee9ade00c683-kube-api-access-zzdp9\") pod \"community-operators-khr6h\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.578489 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qx9d8"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.620896 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.621074 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6mx5\" (UniqueName: \"kubernetes.io/projected/651601bd-18fe-4ca1-9c61-481ca568d022-kube-api-access-n6mx5\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.621124 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-catalog-content\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.621156 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-utilities\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.621324 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.121307765 +0000 UTC m=+192.570003161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.692394 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.701716 4778 generic.go:334] "Generic (PLEG): container finished" podID="a240fd7b-5854-4548-a847-e5590111964b" containerID="6e9a4135f2199a3918c9a565e1055b2ed771be6904f7c3aed074108524b55a58" exitCode=0 Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.701808 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" event={"ID":"a240fd7b-5854-4548-a847-e5590111964b","Type":"ContainerDied","Data":"6e9a4135f2199a3918c9a565e1055b2ed771be6904f7c3aed074108524b55a58"} Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.704696 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sjk9p"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.705531 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.721237 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sjk9p"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.721945 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6mx5\" (UniqueName: \"kubernetes.io/projected/651601bd-18fe-4ca1-9c61-481ca568d022-kube-api-access-n6mx5\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.722012 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-catalog-content\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.722045 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-utilities\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.722071 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.722378 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.22236379 +0000 UTC m=+192.671059186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.722673 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-catalog-content\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.722747 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-utilities\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.739714 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" event={"ID":"2be5b8df-aaff-4a2b-9b54-78a7e58bc420","Type":"ContainerStarted","Data":"c6196405a97553844994349bb1122565b124ed98a010d4bd3c296e91e0e396bc"} Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.766341 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6mx5\" (UniqueName: \"kubernetes.io/projected/651601bd-18fe-4ca1-9c61-481ca568d022-kube-api-access-n6mx5\") pod \"certified-operators-qx9d8\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.824714 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.825093 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq7jg\" (UniqueName: \"kubernetes.io/projected/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-kube-api-access-dq7jg\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.825153 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-catalog-content\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.825282 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-utilities\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.825876 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.325861544 +0000 UTC m=+192.774556940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.862587 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.910303 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l8n9b"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.922309 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.928877 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq7jg\" (UniqueName: \"kubernetes.io/projected/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-kube-api-access-dq7jg\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.928922 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-catalog-content\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.928965 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-utilities\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.929021 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:53 crc kubenswrapper[4778]: E0312 13:12:53.929309 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.429295185 +0000 UTC m=+192.877990581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.929374 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l8n9b"] Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.929954 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-catalog-content\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.930170 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-utilities\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:53 crc kubenswrapper[4778]: I0312 13:12:53.996791 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq7jg\" (UniqueName: \"kubernetes.io/projected/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-kube-api-access-dq7jg\") pod \"community-operators-sjk9p\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.019026 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.034784 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.034983 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-utilities\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.035036 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-catalog-content\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.035083 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxz76\" (UniqueName: \"kubernetes.io/projected/c27afe2a-3402-49f9-b985-45fe67e40d22-kube-api-access-dxz76\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: E0312 13:12:54.035202 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.535171246 +0000 UTC m=+192.983866642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.063910 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4bxj" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.141079 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.141415 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxz76\" (UniqueName: \"kubernetes.io/projected/c27afe2a-3402-49f9-b985-45fe67e40d22-kube-api-access-dxz76\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.141535 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-utilities\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.141574 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-catalog-content\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.142032 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-catalog-content\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: E0312 13:12:54.142328 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.642313632 +0000 UTC m=+193.091009028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.142978 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-utilities\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.177405 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-khr6h"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.178579 4778 ???:1] "http: TLS handshake error from 192.168.126.11:50664: no serving certificate available for the kubelet" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.180811 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxz76\" (UniqueName: \"kubernetes.io/projected/c27afe2a-3402-49f9-b985-45fe67e40d22-kube-api-access-dxz76\") pod \"certified-operators-l8n9b\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: W0312 13:12:54.191323 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d185732_cd6b_44c6_b4db_ee9ade00c683.slice/crio-f4257f2b5ae0b8d1695cb20eed3d7af4ca3c14b5f906e52fd4e46f8237158ff5 WatchSource:0}: Error finding container f4257f2b5ae0b8d1695cb20eed3d7af4ca3c14b5f906e52fd4e46f8237158ff5: Status 404 returned error can't find the container with id f4257f2b5ae0b8d1695cb20eed3d7af4ca3c14b5f906e52fd4e46f8237158ff5 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.245009 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:54 crc kubenswrapper[4778]: E0312 13:12:54.245456 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.745440935 +0000 UTC m=+193.194136331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.284693 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pgrb5"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.284940 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerName="controller-manager" containerID="cri-o://1d59a8a49996a8759d79ab136fc846582c037b05e8fb37ea5f281de757e513e2" gracePeriod=30 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.308553 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.326109 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.327653 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerName="route-controller-manager" containerID="cri-o://c400bf292471252407338cf73137e8439d1cb8b7e278bdf9b5b3d6aae90e459c" gracePeriod=30 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.347703 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: E0312 13:12:54.348068 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.848056764 +0000 UTC m=+193.296752160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.447581 4778 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.454775 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:54 crc kubenswrapper[4778]: E0312 13:12:54.454919 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.954898511 +0000 UTC m=+193.403593907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:54 crc kubenswrapper[4778]: E0312 13:12:54.455382 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 13:12:54.955374285 +0000 UTC m=+193.404069671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxrx4" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.456267 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.461411 4778 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-12T13:12:54.447835992Z","Handler":null,"Name":""} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.465419 4778 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.465452 4778 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.502049 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qx9d8"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.509857 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:54 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:54 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:54 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.509897 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.558080 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.570841 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.586732 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sjk9p"] Mar 12 13:12:54 crc kubenswrapper[4778]: W0312 13:12:54.605315 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b3fb69e_dd4f_4787_a207_4fe25106f9e7.slice/crio-54d14e24e2014de0b1846a5aa684b84b3bf2783c8e0d47fb26e64cb9f10b0a8d WatchSource:0}: Error finding container 54d14e24e2014de0b1846a5aa684b84b3bf2783c8e0d47fb26e64cb9f10b0a8d: Status 404 returned error can't find the container with id 54d14e24e2014de0b1846a5aa684b84b3bf2783c8e0d47fb26e64cb9f10b0a8d Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.659931 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.665596 4778 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.665626 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.763108 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjk9p" event={"ID":"3b3fb69e-dd4f-4787-a207-4fe25106f9e7","Type":"ContainerStarted","Data":"54d14e24e2014de0b1846a5aa684b84b3bf2783c8e0d47fb26e64cb9f10b0a8d"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.770633 4778 generic.go:334] "Generic (PLEG): container finished" podID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerID="05d961ad3b7bd74a33e24a693f2775dd8f5c4483b25df2fe323f0e88cb5ff934" exitCode=0 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.770851 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khr6h" event={"ID":"1d185732-cd6b-44c6-b4db-ee9ade00c683","Type":"ContainerDied","Data":"05d961ad3b7bd74a33e24a693f2775dd8f5c4483b25df2fe323f0e88cb5ff934"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.770916 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khr6h" event={"ID":"1d185732-cd6b-44c6-b4db-ee9ade00c683","Type":"ContainerStarted","Data":"f4257f2b5ae0b8d1695cb20eed3d7af4ca3c14b5f906e52fd4e46f8237158ff5"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.775870 4778 generic.go:334] "Generic (PLEG): container finished" podID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerID="c400bf292471252407338cf73137e8439d1cb8b7e278bdf9b5b3d6aae90e459c" exitCode=0 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.775929 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" event={"ID":"06ffdff1-2f10-4f38-b7fd-b98e883bbc63","Type":"ContainerDied","Data":"c400bf292471252407338cf73137e8439d1cb8b7e278bdf9b5b3d6aae90e459c"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.779877 4778 generic.go:334] "Generic (PLEG): container finished" podID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerID="1d59a8a49996a8759d79ab136fc846582c037b05e8fb37ea5f281de757e513e2" exitCode=0 Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.779935 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" event={"ID":"06bbf7b7-3e40-4aa0-a3db-a56897f5488c","Type":"ContainerDied","Data":"1d59a8a49996a8759d79ab136fc846582c037b05e8fb37ea5f281de757e513e2"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.784665 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" event={"ID":"2be5b8df-aaff-4a2b-9b54-78a7e58bc420","Type":"ContainerStarted","Data":"7bf4ff3391edb99ce741e5397b58436a4780a99f8d39dcd8b173aff2b9bac9fa"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.786329 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qx9d8" event={"ID":"651601bd-18fe-4ca1-9c61-481ca568d022","Type":"ContainerStarted","Data":"592ed663fa0a363547ba9675a7740b1982ac31820675fa1bc6b541164ee13dff"} Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.790804 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxrx4\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.833171 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.833781 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.843724 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.844018 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.873583 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.910079 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.913848 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.957868 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l8n9b"] Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.962263 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.965392 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-config\") pod \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.965447 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k56zt\" (UniqueName: \"kubernetes.io/projected/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-kube-api-access-k56zt\") pod \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.965469 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-serving-cert\") pod \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.965520 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-client-ca\") pod \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\" (UID: \"06ffdff1-2f10-4f38-b7fd-b98e883bbc63\") " Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.965649 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/593bf507-3097-460c-aa84-c680a76f3ffe-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.965666 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/593bf507-3097-460c-aa84-c680a76f3ffe-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.966472 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-config" (OuterVolumeSpecName: "config") pod "06ffdff1-2f10-4f38-b7fd-b98e883bbc63" (UID: "06ffdff1-2f10-4f38-b7fd-b98e883bbc63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.966999 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-client-ca" (OuterVolumeSpecName: "client-ca") pod "06ffdff1-2f10-4f38-b7fd-b98e883bbc63" (UID: "06ffdff1-2f10-4f38-b7fd-b98e883bbc63"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.972524 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-kube-api-access-k56zt" (OuterVolumeSpecName: "kube-api-access-k56zt") pod "06ffdff1-2f10-4f38-b7fd-b98e883bbc63" (UID: "06ffdff1-2f10-4f38-b7fd-b98e883bbc63"). InnerVolumeSpecName "kube-api-access-k56zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:12:54 crc kubenswrapper[4778]: I0312 13:12:54.979400 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "06ffdff1-2f10-4f38-b7fd-b98e883bbc63" (UID: "06ffdff1-2f10-4f38-b7fd-b98e883bbc63"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.066344 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-serving-cert\") pod \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.066643 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-client-ca\") pod \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.066715 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-proxy-ca-bundles\") pod \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.066737 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9bl9\" (UniqueName: \"kubernetes.io/projected/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-kube-api-access-c9bl9\") pod \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.066761 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-config\") pod \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\" (UID: \"06bbf7b7-3e40-4aa0-a3db-a56897f5488c\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.067434 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "06bbf7b7-3e40-4aa0-a3db-a56897f5488c" (UID: "06bbf7b7-3e40-4aa0-a3db-a56897f5488c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.067631 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/593bf507-3097-460c-aa84-c680a76f3ffe-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.067655 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/593bf507-3097-460c-aa84-c680a76f3ffe-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.067745 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-client-ca" (OuterVolumeSpecName: "client-ca") pod "06bbf7b7-3e40-4aa0-a3db-a56897f5488c" (UID: "06bbf7b7-3e40-4aa0-a3db-a56897f5488c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.069909 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "06bbf7b7-3e40-4aa0-a3db-a56897f5488c" (UID: "06bbf7b7-3e40-4aa0-a3db-a56897f5488c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.069980 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-config" (OuterVolumeSpecName: "config") pod "06bbf7b7-3e40-4aa0-a3db-a56897f5488c" (UID: "06bbf7b7-3e40-4aa0-a3db-a56897f5488c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070364 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070525 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070533 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-kube-api-access-c9bl9" (OuterVolumeSpecName: "kube-api-access-c9bl9") pod "06bbf7b7-3e40-4aa0-a3db-a56897f5488c" (UID: "06bbf7b7-3e40-4aa0-a3db-a56897f5488c"). InnerVolumeSpecName "kube-api-access-c9bl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070549 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k56zt\" (UniqueName: \"kubernetes.io/projected/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-kube-api-access-k56zt\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070568 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ffdff1-2f10-4f38-b7fd-b98e883bbc63-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070580 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.070469 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/593bf507-3097-460c-aa84-c680a76f3ffe-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.073072 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.087210 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/593bf507-3097-460c-aa84-c680a76f3ffe-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.171601 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a240fd7b-5854-4548-a847-e5590111964b-secret-volume\") pod \"a240fd7b-5854-4548-a847-e5590111964b\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.171656 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88x7r\" (UniqueName: \"kubernetes.io/projected/a240fd7b-5854-4548-a847-e5590111964b-kube-api-access-88x7r\") pod \"a240fd7b-5854-4548-a847-e5590111964b\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.171691 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume\") pod \"a240fd7b-5854-4548-a847-e5590111964b\" (UID: \"a240fd7b-5854-4548-a847-e5590111964b\") " Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.171993 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9bl9\" (UniqueName: \"kubernetes.io/projected/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-kube-api-access-c9bl9\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.172010 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.172018 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.172026 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06bbf7b7-3e40-4aa0-a3db-a56897f5488c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.172425 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume" (OuterVolumeSpecName: "config-volume") pod "a240fd7b-5854-4548-a847-e5590111964b" (UID: "a240fd7b-5854-4548-a847-e5590111964b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.176539 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a240fd7b-5854-4548-a847-e5590111964b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a240fd7b-5854-4548-a847-e5590111964b" (UID: "a240fd7b-5854-4548-a847-e5590111964b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.182311 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a240fd7b-5854-4548-a847-e5590111964b-kube-api-access-88x7r" (OuterVolumeSpecName: "kube-api-access-88x7r") pod "a240fd7b-5854-4548-a847-e5590111964b" (UID: "a240fd7b-5854-4548-a847-e5590111964b"). InnerVolumeSpecName "kube-api-access-88x7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.192975 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.206995 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm"] Mar 12 13:12:55 crc kubenswrapper[4778]: E0312 13:12:55.207236 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerName="controller-manager" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.207250 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerName="controller-manager" Mar 12 13:12:55 crc kubenswrapper[4778]: E0312 13:12:55.207267 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerName="route-controller-manager" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.207277 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerName="route-controller-manager" Mar 12 13:12:55 crc kubenswrapper[4778]: E0312 13:12:55.207288 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a240fd7b-5854-4548-a847-e5590111964b" containerName="collect-profiles" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.207296 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a240fd7b-5854-4548-a847-e5590111964b" containerName="collect-profiles" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.207414 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" containerName="controller-manager" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.207433 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a240fd7b-5854-4548-a847-e5590111964b" containerName="collect-profiles" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.207450 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" containerName="route-controller-manager" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.208218 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.220471 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.226980 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.227720 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.232057 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273630 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lwnl\" (UniqueName: \"kubernetes.io/projected/b330900c-c52a-4e88-a2d2-38e34f837004-kube-api-access-4lwnl\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273675 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-client-ca\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273708 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-client-ca\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273724 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-proxy-ca-bundles\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273776 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qnmc\" (UniqueName: \"kubernetes.io/projected/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-kube-api-access-2qnmc\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273794 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-config\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273814 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-config\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273843 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b330900c-c52a-4e88-a2d2-38e34f837004-serving-cert\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273862 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-serving-cert\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273903 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a240fd7b-5854-4548-a847-e5590111964b-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273913 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88x7r\" (UniqueName: \"kubernetes.io/projected/a240fd7b-5854-4548-a847-e5590111964b-kube-api-access-88x7r\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.273924 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a240fd7b-5854-4548-a847-e5590111964b-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.304781 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8xksl"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.305743 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.311624 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.322363 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xksl"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376085 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-config\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376163 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b330900c-c52a-4e88-a2d2-38e34f837004-serving-cert\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376212 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-utilities\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376237 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-serving-cert\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376281 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lwnl\" (UniqueName: \"kubernetes.io/projected/b330900c-c52a-4e88-a2d2-38e34f837004-kube-api-access-4lwnl\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376313 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-client-ca\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376336 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kch8z\" (UniqueName: \"kubernetes.io/projected/de4557b4-7957-47a0-8c42-845be1fa0f32-kube-api-access-kch8z\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376373 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-client-ca\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376403 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-proxy-ca-bundles\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376443 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-catalog-content\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376480 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qnmc\" (UniqueName: \"kubernetes.io/projected/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-kube-api-access-2qnmc\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.376505 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-config\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.378507 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-config\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.379090 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-config\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.382751 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b330900c-c52a-4e88-a2d2-38e34f837004-serving-cert\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.383438 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-client-ca\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.384233 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-serving-cert\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.384336 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-proxy-ca-bundles\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.385073 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-client-ca\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.401741 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxrx4"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.407947 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lwnl\" (UniqueName: \"kubernetes.io/projected/b330900c-c52a-4e88-a2d2-38e34f837004-kube-api-access-4lwnl\") pod \"controller-manager-7b56f5b6c6-7q5xm\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.408854 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qnmc\" (UniqueName: \"kubernetes.io/projected/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-kube-api-access-2qnmc\") pod \"route-controller-manager-5f6cfcbfb9-jcsqb\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: W0312 13:12:55.436466 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51ee714f_fb23_4420_9e70_1b3134eea18e.slice/crio-9fdf9bc3368d582f75afb64ef1bd7b59c9e3cd5fe63a9b2265425474dba3a3b4 WatchSource:0}: Error finding container 9fdf9bc3368d582f75afb64ef1bd7b59c9e3cd5fe63a9b2265425474dba3a3b4: Status 404 returned error can't find the container with id 9fdf9bc3368d582f75afb64ef1bd7b59c9e3cd5fe63a9b2265425474dba3a3b4 Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.477883 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-catalog-content\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.477954 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-utilities\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.477994 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kch8z\" (UniqueName: \"kubernetes.io/projected/de4557b4-7957-47a0-8c42-845be1fa0f32-kube-api-access-kch8z\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.479059 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-utilities\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.479084 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-catalog-content\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.493026 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kch8z\" (UniqueName: \"kubernetes.io/projected/de4557b4-7957-47a0-8c42-845be1fa0f32-kube-api-access-kch8z\") pod \"redhat-marketplace-8xksl\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.509048 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:55 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:55 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:55 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.509114 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.563606 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.577137 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.633534 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.697915 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rtjz5"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.698832 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.709410 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtjz5"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.719617 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.784263 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-catalog-content\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.784372 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjm7s\" (UniqueName: \"kubernetes.io/projected/b9bef112-9bef-4ce2-abd8-054b4d671658-kube-api-access-gjm7s\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.784439 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-utilities\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.815274 4778 generic.go:334] "Generic (PLEG): container finished" podID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerID="beac9341cf9caf9b2899c0d3555998167e4413386821c255145cfe1b113c1402" exitCode=0 Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.815412 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l8n9b" event={"ID":"c27afe2a-3402-49f9-b985-45fe67e40d22","Type":"ContainerDied","Data":"beac9341cf9caf9b2899c0d3555998167e4413386821c255145cfe1b113c1402"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.815549 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l8n9b" event={"ID":"c27afe2a-3402-49f9-b985-45fe67e40d22","Type":"ContainerStarted","Data":"f3e464dc52992fdb0f0b53c632c09c98afcb767da1a2f76ffc34b25c53dcb6a3"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.817379 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" event={"ID":"06bbf7b7-3e40-4aa0-a3db-a56897f5488c","Type":"ContainerDied","Data":"baecc290d5904f2078cb76008ee3fad41b6baea1393aa1ce14dba9ed727aca24"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.817405 4778 scope.go:117] "RemoveContainer" containerID="1d59a8a49996a8759d79ab136fc846582c037b05e8fb37ea5f281de757e513e2" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.817486 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pgrb5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.838998 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.857766 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pgrb5"] Mar 12 13:12:55 crc kubenswrapper[4778]: W0312 13:12:55.862211 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb330900c_c52a_4e88_a2d2_38e34f837004.slice/crio-17004ca7182405d5ca539723da4c058b3a73eeda64b060c4ba9deebe55691d77 WatchSource:0}: Error finding container 17004ca7182405d5ca539723da4c058b3a73eeda64b060c4ba9deebe55691d77: Status 404 returned error can't find the container with id 17004ca7182405d5ca539723da4c058b3a73eeda64b060c4ba9deebe55691d77 Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.874136 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pgrb5"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.875106 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" event={"ID":"2be5b8df-aaff-4a2b-9b54-78a7e58bc420","Type":"ContainerStarted","Data":"4810a5582bbaa1f78751f17e1a82841738eba7c02be4b8158216a27251deac87"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.886522 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-catalog-content\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.886611 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjm7s\" (UniqueName: \"kubernetes.io/projected/b9bef112-9bef-4ce2-abd8-054b4d671658-kube-api-access-gjm7s\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.886692 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-utilities\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.887113 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-utilities\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.887413 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-catalog-content\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.892498 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjk9p" event={"ID":"3b3fb69e-dd4f-4787-a207-4fe25106f9e7","Type":"ContainerDied","Data":"abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.892036 4778 generic.go:334] "Generic (PLEG): container finished" podID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerID="abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961" exitCode=0 Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.898559 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" event={"ID":"51ee714f-fb23-4420-9e70-1b3134eea18e","Type":"ContainerStarted","Data":"29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.898600 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" event={"ID":"51ee714f-fb23-4420-9e70-1b3134eea18e","Type":"ContainerStarted","Data":"9fdf9bc3368d582f75afb64ef1bd7b59c9e3cd5fe63a9b2265425474dba3a3b4"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.898791 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.919001 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.919035 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"593bf507-3097-460c-aa84-c680a76f3ffe","Type":"ContainerStarted","Data":"b74dd280126159ecf287aae9295c5a458084d9b654a5ee2ddd9fbd12aaafb12b"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.926081 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjm7s\" (UniqueName: \"kubernetes.io/projected/b9bef112-9bef-4ce2-abd8-054b4d671658-kube-api-access-gjm7s\") pod \"redhat-marketplace-rtjz5\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.949997 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cp2lw" podStartSLOduration=11.949842583 podStartE2EDuration="11.949842583s" podCreationTimestamp="2026-03-12 13:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:55.918848181 +0000 UTC m=+194.367543577" watchObservedRunningTime="2026-03-12 13:12:55.949842583 +0000 UTC m=+194.398537979" Mar 12 13:12:55 crc kubenswrapper[4778]: W0312 13:12:55.960409 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fa0e405_9e9d_49fc_b2aa_17ca5c529a74.slice/crio-d26df510f14aa09fdd3bb4f8a1ddd8e15d52996d36167c7af13ebde7af2c80d1 WatchSource:0}: Error finding container d26df510f14aa09fdd3bb4f8a1ddd8e15d52996d36167c7af13ebde7af2c80d1: Status 404 returned error can't find the container with id d26df510f14aa09fdd3bb4f8a1ddd8e15d52996d36167c7af13ebde7af2c80d1 Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.960536 4778 generic.go:334] "Generic (PLEG): container finished" podID="651601bd-18fe-4ca1-9c61-481ca568d022" containerID="768c08538cc35f7dca92094b0ee56f8d00acc523e23bc32165393cb6d17f7cd2" exitCode=0 Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.960708 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qx9d8" event={"ID":"651601bd-18fe-4ca1-9c61-481ca568d022","Type":"ContainerDied","Data":"768c08538cc35f7dca92094b0ee56f8d00acc523e23bc32165393cb6d17f7cd2"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.965376 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xksl"] Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.970751 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" podStartSLOduration=144.970733311 podStartE2EDuration="2m24.970733311s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:55.970510945 +0000 UTC m=+194.419206351" watchObservedRunningTime="2026-03-12 13:12:55.970733311 +0000 UTC m=+194.419428707" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.972662 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" event={"ID":"06ffdff1-2f10-4f38-b7fd-b98e883bbc63","Type":"ContainerDied","Data":"58650ee0315d5aac50c162f5420d39a44557cb90a0d565bd9b299a8e4ee0251d"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.972722 4778 scope.go:117] "RemoveContainer" containerID="c400bf292471252407338cf73137e8439d1cb8b7e278bdf9b5b3d6aae90e459c" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.972823 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.979927 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" event={"ID":"a240fd7b-5854-4548-a847-e5590111964b","Type":"ContainerDied","Data":"aa7b81ba2e81ec0ae29a489fd430d937b62e78b889977bd807a40c2a99fb3190"} Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.979964 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa7b81ba2e81ec0ae29a489fd430d937b62e78b889977bd807a40c2a99fb3190" Mar 12 13:12:55 crc kubenswrapper[4778]: I0312 13:12:55.979972 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.023428 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.051222 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh"] Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.069986 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zpgxh"] Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.268742 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06bbf7b7-3e40-4aa0-a3db-a56897f5488c" path="/var/lib/kubelet/pods/06bbf7b7-3e40-4aa0-a3db-a56897f5488c/volumes" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.271105 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06ffdff1-2f10-4f38-b7fd-b98e883bbc63" path="/var/lib/kubelet/pods/06ffdff1-2f10-4f38-b7fd-b98e883bbc63/volumes" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.273999 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.278325 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtjz5"] Mar 12 13:12:56 crc kubenswrapper[4778]: W0312 13:12:56.286155 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9bef112_9bef_4ce2_abd8_054b4d671658.slice/crio-823af2a7e3b6063a4f30d49b66161c625efcb36bf067f9d539324e41889ea011 WatchSource:0}: Error finding container 823af2a7e3b6063a4f30d49b66161c625efcb36bf067f9d539324e41889ea011: Status 404 returned error can't find the container with id 823af2a7e3b6063a4f30d49b66161c625efcb36bf067f9d539324e41889ea011 Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.509177 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:56 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:56 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:56 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.509705 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.690420 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5s5vs"] Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.691513 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.694587 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.738852 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5s5vs"] Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.781604 4778 ???:1] "http: TLS handshake error from 192.168.126.11:50680: no serving certificate available for the kubelet" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.799530 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-utilities\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.799573 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-catalog-content\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.799869 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpfz2\" (UniqueName: \"kubernetes.io/projected/f438f2a3-60c0-4554-a49b-030545f8139c-kube-api-access-mpfz2\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.901364 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-catalog-content\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.901494 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpfz2\" (UniqueName: \"kubernetes.io/projected/f438f2a3-60c0-4554-a49b-030545f8139c-kube-api-access-mpfz2\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.901537 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-utilities\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.902384 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-utilities\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.902473 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-catalog-content\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.921386 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpfz2\" (UniqueName: \"kubernetes.io/projected/f438f2a3-60c0-4554-a49b-030545f8139c-kube-api-access-mpfz2\") pod \"redhat-operators-5s5vs\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.992428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" event={"ID":"b330900c-c52a-4e88-a2d2-38e34f837004","Type":"ContainerStarted","Data":"af0097b4c8ffcf21c4d0f3d542c30c13b992c9bb5a36537354858b4fa3539991"} Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.992483 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" event={"ID":"b330900c-c52a-4e88-a2d2-38e34f837004","Type":"ContainerStarted","Data":"17004ca7182405d5ca539723da4c058b3a73eeda64b060c4ba9deebe55691d77"} Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.992895 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.998002 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.998267 4778 generic.go:334] "Generic (PLEG): container finished" podID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerID="167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252" exitCode=0 Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.998315 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xksl" event={"ID":"de4557b4-7957-47a0-8c42-845be1fa0f32","Type":"ContainerDied","Data":"167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252"} Mar 12 13:12:56 crc kubenswrapper[4778]: I0312 13:12:56.998344 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xksl" event={"ID":"de4557b4-7957-47a0-8c42-845be1fa0f32","Type":"ContainerStarted","Data":"775a67dbf14a4aa00ee320f14ee688f2689c34e66ee23b796f0166af1618f55f"} Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.001806 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" event={"ID":"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74","Type":"ContainerStarted","Data":"e1481cb1e8a9421818bd64f72eed1fe038b53c1df39f76f7438a261e48a535ff"} Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.001849 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" event={"ID":"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74","Type":"ContainerStarted","Data":"d26df510f14aa09fdd3bb4f8a1ddd8e15d52996d36167c7af13ebde7af2c80d1"} Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.002384 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.007138 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.007472 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.013223 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" podStartSLOduration=2.013194977 podStartE2EDuration="2.013194977s" podCreationTimestamp="2026-03-12 13:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:57.011311934 +0000 UTC m=+195.460007350" watchObservedRunningTime="2026-03-12 13:12:57.013194977 +0000 UTC m=+195.461890373" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.016707 4778 generic.go:334] "Generic (PLEG): container finished" podID="593bf507-3097-460c-aa84-c680a76f3ffe" containerID="4b23771e05a4ff737085d226f6bad113bb6e4e0f0e89dfb0d1486ecc9eab431e" exitCode=0 Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.016760 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"593bf507-3097-460c-aa84-c680a76f3ffe","Type":"ContainerDied","Data":"4b23771e05a4ff737085d226f6bad113bb6e4e0f0e89dfb0d1486ecc9eab431e"} Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.019752 4778 generic.go:334] "Generic (PLEG): container finished" podID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerID="3ee91beb1526d7d2135a66716b66577b22ca3756c6f18236717330ab9060a779" exitCode=0 Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.020047 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtjz5" event={"ID":"b9bef112-9bef-4ce2-abd8-054b4d671658","Type":"ContainerDied","Data":"3ee91beb1526d7d2135a66716b66577b22ca3756c6f18236717330ab9060a779"} Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.020078 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtjz5" event={"ID":"b9bef112-9bef-4ce2-abd8-054b4d671658","Type":"ContainerStarted","Data":"823af2a7e3b6063a4f30d49b66161c625efcb36bf067f9d539324e41889ea011"} Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.066997 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.067046 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.087272 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.104618 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" podStartSLOduration=2.1045992 podStartE2EDuration="2.1045992s" podCreationTimestamp="2026-03-12 13:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:57.101426051 +0000 UTC m=+195.550121447" watchObservedRunningTime="2026-03-12 13:12:57.1045992 +0000 UTC m=+195.553294596" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.138726 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-76s88"] Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.141125 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.146479 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-76s88"] Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.196110 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.209591 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vnndl" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.214402 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.215862 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.222701 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.223171 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.241065 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.241102 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.242540 4778 patch_prober.go:28] interesting pod/console-f9d7485db-xwwxp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.242572 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xwwxp" podUID="c825022c-79bc-44ae-bc64-ee9614aafe25" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.250755 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.307831 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dc25eb7-f12c-4445-bd35-107ac0c35429-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.307951 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6lb\" (UniqueName: \"kubernetes.io/projected/34ecd758-517c-455a-939a-7eb6d3546854-kube-api-access-cg6lb\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.308001 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-catalog-content\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.308071 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-utilities\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.308092 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dc25eb7-f12c-4445-bd35-107ac0c35429-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.400438 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5s5vs"] Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.412084 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6lb\" (UniqueName: \"kubernetes.io/projected/34ecd758-517c-455a-939a-7eb6d3546854-kube-api-access-cg6lb\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.412155 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-catalog-content\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.412224 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-utilities\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.412251 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dc25eb7-f12c-4445-bd35-107ac0c35429-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.412296 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dc25eb7-f12c-4445-bd35-107ac0c35429-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.413458 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-utilities\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.413487 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-catalog-content\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.413523 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dc25eb7-f12c-4445-bd35-107ac0c35429-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.436320 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dc25eb7-f12c-4445-bd35-107ac0c35429-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.438328 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6lb\" (UniqueName: \"kubernetes.io/projected/34ecd758-517c-455a-939a-7eb6d3546854-kube-api-access-cg6lb\") pod \"redhat-operators-76s88\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.479797 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.479864 4778 patch_prober.go:28] interesting pod/downloads-7954f5f757-mx6kn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.479876 4778 patch_prober.go:28] interesting pod/downloads-7954f5f757-mx6kn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.479892 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mx6kn" podUID="8af48f77-25f7-49ca-8bcb-2481aa72ee66" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.479923 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mx6kn" podUID="8af48f77-25f7-49ca-8bcb-2481aa72ee66" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.504418 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.508361 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:57 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:57 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:57 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.508395 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.575423 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.843378 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-76s88"] Mar 12 13:12:57 crc kubenswrapper[4778]: W0312 13:12:57.850983 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34ecd758_517c_455a_939a_7eb6d3546854.slice/crio-cf68cb478854e264cd59c9ad8e9f3e763498e2e2706254a3b88fc3dd9f22fe4f WatchSource:0}: Error finding container cf68cb478854e264cd59c9ad8e9f3e763498e2e2706254a3b88fc3dd9f22fe4f: Status 404 returned error can't find the container with id cf68cb478854e264cd59c9ad8e9f3e763498e2e2706254a3b88fc3dd9f22fe4f Mar 12 13:12:57 crc kubenswrapper[4778]: I0312 13:12:57.948439 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 13:12:57 crc kubenswrapper[4778]: W0312 13:12:57.956836 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5dc25eb7_f12c_4445_bd35_107ac0c35429.slice/crio-877b25ba26cb3a576a6780cb185f284328b74ec31202ca990242a090aee0ba1f WatchSource:0}: Error finding container 877b25ba26cb3a576a6780cb185f284328b74ec31202ca990242a090aee0ba1f: Status 404 returned error can't find the container with id 877b25ba26cb3a576a6780cb185f284328b74ec31202ca990242a090aee0ba1f Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.032923 4778 generic.go:334] "Generic (PLEG): container finished" podID="f438f2a3-60c0-4554-a49b-030545f8139c" containerID="9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e" exitCode=0 Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.033088 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerDied","Data":"9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e"} Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.033329 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerStarted","Data":"c5e7e785f566d6c012fb07b0778c4b6c15691ef04836f8607417e605e9c6feb5"} Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.039230 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerStarted","Data":"5ab6ab1e87e3d9a4f7941a7ab56868950f541c7821fdd08fb7b7e95206f0cb25"} Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.039267 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerStarted","Data":"cf68cb478854e264cd59c9ad8e9f3e763498e2e2706254a3b88fc3dd9f22fe4f"} Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.042741 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5dc25eb7-f12c-4445-bd35-107ac0c35429","Type":"ContainerStarted","Data":"877b25ba26cb3a576a6780cb185f284328b74ec31202ca990242a090aee0ba1f"} Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.049891 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-xz42x" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.353965 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.507891 4778 patch_prober.go:28] interesting pod/router-default-5444994796-ms5xq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 13:12:58 crc kubenswrapper[4778]: [-]has-synced failed: reason withheld Mar 12 13:12:58 crc kubenswrapper[4778]: [+]process-running ok Mar 12 13:12:58 crc kubenswrapper[4778]: healthz check failed Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.507951 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ms5xq" podUID="5c8d947a-b62b-4eb9-81d7-94530285e8dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.546717 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/593bf507-3097-460c-aa84-c680a76f3ffe-kube-api-access\") pod \"593bf507-3097-460c-aa84-c680a76f3ffe\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.546813 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/593bf507-3097-460c-aa84-c680a76f3ffe-kubelet-dir\") pod \"593bf507-3097-460c-aa84-c680a76f3ffe\" (UID: \"593bf507-3097-460c-aa84-c680a76f3ffe\") " Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.547243 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/593bf507-3097-460c-aa84-c680a76f3ffe-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "593bf507-3097-460c-aa84-c680a76f3ffe" (UID: "593bf507-3097-460c-aa84-c680a76f3ffe"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.557402 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.557471 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.562430 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593bf507-3097-460c-aa84-c680a76f3ffe-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "593bf507-3097-460c-aa84-c680a76f3ffe" (UID: "593bf507-3097-460c-aa84-c680a76f3ffe"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.648951 4778 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/593bf507-3097-460c-aa84-c680a76f3ffe-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.648980 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/593bf507-3097-460c-aa84-c680a76f3ffe-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:12:58 crc kubenswrapper[4778]: I0312 13:12:58.661958 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.058911 4778 generic.go:334] "Generic (PLEG): container finished" podID="34ecd758-517c-455a-939a-7eb6d3546854" containerID="5ab6ab1e87e3d9a4f7941a7ab56868950f541c7821fdd08fb7b7e95206f0cb25" exitCode=0 Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.059200 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerDied","Data":"5ab6ab1e87e3d9a4f7941a7ab56868950f541c7821fdd08fb7b7e95206f0cb25"} Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.062017 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5dc25eb7-f12c-4445-bd35-107ac0c35429","Type":"ContainerStarted","Data":"6fd9810922229212229102ca032fc9237ac433c61fc8c7136009fc1f0bbe286e"} Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.063908 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"593bf507-3097-460c-aa84-c680a76f3ffe","Type":"ContainerDied","Data":"b74dd280126159ecf287aae9295c5a458084d9b654a5ee2ddd9fbd12aaafb12b"} Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.063953 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b74dd280126159ecf287aae9295c5a458084d9b654a5ee2ddd9fbd12aaafb12b" Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.064080 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.509682 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.514784 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ms5xq" Mar 12 13:12:59 crc kubenswrapper[4778]: I0312 13:12:59.547347 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.547321391 podStartE2EDuration="2.547321391s" podCreationTimestamp="2026-03-12 13:12:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:12:59.087080389 +0000 UTC m=+197.535775785" watchObservedRunningTime="2026-03-12 13:12:59.547321391 +0000 UTC m=+197.996016787" Mar 12 13:13:00 crc kubenswrapper[4778]: I0312 13:13:00.088333 4778 generic.go:334] "Generic (PLEG): container finished" podID="5dc25eb7-f12c-4445-bd35-107ac0c35429" containerID="6fd9810922229212229102ca032fc9237ac433c61fc8c7136009fc1f0bbe286e" exitCode=0 Mar 12 13:13:00 crc kubenswrapper[4778]: I0312 13:13:00.088386 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5dc25eb7-f12c-4445-bd35-107ac0c35429","Type":"ContainerDied","Data":"6fd9810922229212229102ca032fc9237ac433c61fc8c7136009fc1f0bbe286e"} Mar 12 13:13:00 crc kubenswrapper[4778]: I0312 13:13:00.143972 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-8zmxq" Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.506915 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.607057 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dc25eb7-f12c-4445-bd35-107ac0c35429-kubelet-dir\") pod \"5dc25eb7-f12c-4445-bd35-107ac0c35429\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.607176 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dc25eb7-f12c-4445-bd35-107ac0c35429-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5dc25eb7-f12c-4445-bd35-107ac0c35429" (UID: "5dc25eb7-f12c-4445-bd35-107ac0c35429"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.607289 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dc25eb7-f12c-4445-bd35-107ac0c35429-kube-api-access\") pod \"5dc25eb7-f12c-4445-bd35-107ac0c35429\" (UID: \"5dc25eb7-f12c-4445-bd35-107ac0c35429\") " Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.607641 4778 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dc25eb7-f12c-4445-bd35-107ac0c35429-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.624562 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc25eb7-f12c-4445-bd35-107ac0c35429-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5dc25eb7-f12c-4445-bd35-107ac0c35429" (UID: "5dc25eb7-f12c-4445-bd35-107ac0c35429"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.709039 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dc25eb7-f12c-4445-bd35-107ac0c35429-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:01 crc kubenswrapper[4778]: I0312 13:13:01.920038 4778 ???:1] "http: TLS handshake error from 192.168.126.11:50684: no serving certificate available for the kubelet" Mar 12 13:13:02 crc kubenswrapper[4778]: I0312 13:13:02.054119 4778 ???:1] "http: TLS handshake error from 192.168.126.11:50700: no serving certificate available for the kubelet" Mar 12 13:13:02 crc kubenswrapper[4778]: I0312 13:13:02.122973 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5dc25eb7-f12c-4445-bd35-107ac0c35429","Type":"ContainerDied","Data":"877b25ba26cb3a576a6780cb185f284328b74ec31202ca990242a090aee0ba1f"} Mar 12 13:13:02 crc kubenswrapper[4778]: I0312 13:13:02.123006 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="877b25ba26cb3a576a6780cb185f284328b74ec31202ca990242a090aee0ba1f" Mar 12 13:13:02 crc kubenswrapper[4778]: I0312 13:13:02.123043 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.166994 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.167359 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.167423 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.167455 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.168577 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.169274 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.170378 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.180168 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.183742 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.191234 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.191512 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.192524 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.268347 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.270566 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.275703 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.284764 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0b59b25a-3acc-4d06-b91d-575f45463520-metrics-certs\") pod \"network-metrics-daemon-rz9vw\" (UID: \"0b59b25a-3acc-4d06-b91d-575f45463520\") " pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.288492 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.305548 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.312884 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rz9vw" Mar 12 13:13:06 crc kubenswrapper[4778]: I0312 13:13:06.467090 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 13:13:07 crc kubenswrapper[4778]: I0312 13:13:07.360476 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:13:07 crc kubenswrapper[4778]: I0312 13:13:07.371560 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:13:07 crc kubenswrapper[4778]: I0312 13:13:07.481801 4778 patch_prober.go:28] interesting pod/downloads-7954f5f757-mx6kn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 12 13:13:07 crc kubenswrapper[4778]: I0312 13:13:07.481871 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mx6kn" podUID="8af48f77-25f7-49ca-8bcb-2481aa72ee66" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 12 13:13:07 crc kubenswrapper[4778]: I0312 13:13:07.489423 4778 patch_prober.go:28] interesting pod/downloads-7954f5f757-mx6kn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 12 13:13:07 crc kubenswrapper[4778]: I0312 13:13:07.489485 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mx6kn" podUID="8af48f77-25f7-49ca-8bcb-2481aa72ee66" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 12 13:13:12 crc kubenswrapper[4778]: I0312 13:13:12.184830 4778 ???:1] "http: TLS handshake error from 192.168.126.11:56534: no serving certificate available for the kubelet" Mar 12 13:13:13 crc kubenswrapper[4778]: I0312 13:13:13.831421 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm"] Mar 12 13:13:13 crc kubenswrapper[4778]: I0312 13:13:13.831690 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" containerName="controller-manager" containerID="cri-o://af0097b4c8ffcf21c4d0f3d542c30c13b992c9bb5a36537354858b4fa3539991" gracePeriod=30 Mar 12 13:13:13 crc kubenswrapper[4778]: I0312 13:13:13.841103 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb"] Mar 12 13:13:13 crc kubenswrapper[4778]: I0312 13:13:13.841340 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerName="route-controller-manager" containerID="cri-o://e1481cb1e8a9421818bd64f72eed1fe038b53c1df39f76f7438a261e48a535ff" gracePeriod=30 Mar 12 13:13:14 crc kubenswrapper[4778]: I0312 13:13:14.920385 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.208938 4778 generic.go:334] "Generic (PLEG): container finished" podID="b330900c-c52a-4e88-a2d2-38e34f837004" containerID="af0097b4c8ffcf21c4d0f3d542c30c13b992c9bb5a36537354858b4fa3539991" exitCode=0 Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.209047 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" event={"ID":"b330900c-c52a-4e88-a2d2-38e34f837004","Type":"ContainerDied","Data":"af0097b4c8ffcf21c4d0f3d542c30c13b992c9bb5a36537354858b4fa3539991"} Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.210993 4778 generic.go:334] "Generic (PLEG): container finished" podID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerID="e1481cb1e8a9421818bd64f72eed1fe038b53c1df39f76f7438a261e48a535ff" exitCode=0 Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.211103 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" event={"ID":"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74","Type":"ContainerDied","Data":"e1481cb1e8a9421818bd64f72eed1fe038b53c1df39f76f7438a261e48a535ff"} Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.565509 4778 patch_prober.go:28] interesting pod/controller-manager-7b56f5b6c6-7q5xm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" start-of-body= Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.565577 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.581273 4778 patch_prober.go:28] interesting pod/route-controller-manager-5f6cfcbfb9-jcsqb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" start-of-body= Mar 12 13:13:15 crc kubenswrapper[4778]: I0312 13:13:15.581409 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" Mar 12 13:13:17 crc kubenswrapper[4778]: I0312 13:13:17.485641 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mx6kn" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.172839 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.201903 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w"] Mar 12 13:13:26 crc kubenswrapper[4778]: E0312 13:13:26.202089 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerName="route-controller-manager" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202100 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerName="route-controller-manager" Mar 12 13:13:26 crc kubenswrapper[4778]: E0312 13:13:26.202114 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc25eb7-f12c-4445-bd35-107ac0c35429" containerName="pruner" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202119 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc25eb7-f12c-4445-bd35-107ac0c35429" containerName="pruner" Mar 12 13:13:26 crc kubenswrapper[4778]: E0312 13:13:26.202127 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593bf507-3097-460c-aa84-c680a76f3ffe" containerName="pruner" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202133 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="593bf507-3097-460c-aa84-c680a76f3ffe" containerName="pruner" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202240 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerName="route-controller-manager" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202253 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="593bf507-3097-460c-aa84-c680a76f3ffe" containerName="pruner" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202263 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc25eb7-f12c-4445-bd35-107ac0c35429" containerName="pruner" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.202566 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.207809 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w"] Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.226735 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qnmc\" (UniqueName: \"kubernetes.io/projected/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-kube-api-access-2qnmc\") pod \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227074 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-serving-cert\") pod \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227142 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-client-ca\") pod \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227203 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-config\") pod \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\" (UID: \"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74\") " Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227364 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85126e1b-2b92-4e55-b847-d55f8b1b387e-serving-cert\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227415 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6452d\" (UniqueName: \"kubernetes.io/projected/85126e1b-2b92-4e55-b847-d55f8b1b387e-kube-api-access-6452d\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227472 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-config\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.227498 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-client-ca\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.228149 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-client-ca" (OuterVolumeSpecName: "client-ca") pod "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" (UID: "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.228286 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-config" (OuterVolumeSpecName: "config") pod "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" (UID: "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.233667 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-kube-api-access-2qnmc" (OuterVolumeSpecName: "kube-api-access-2qnmc") pod "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" (UID: "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74"). InnerVolumeSpecName "kube-api-access-2qnmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.235540 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" (UID: "1fa0e405-9e9d-49fc-b2aa-17ca5c529a74"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.275537 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" event={"ID":"1fa0e405-9e9d-49fc-b2aa-17ca5c529a74","Type":"ContainerDied","Data":"d26df510f14aa09fdd3bb4f8a1ddd8e15d52996d36167c7af13ebde7af2c80d1"} Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.275586 4778 scope.go:117] "RemoveContainer" containerID="e1481cb1e8a9421818bd64f72eed1fe038b53c1df39f76f7438a261e48a535ff" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.275633 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.298804 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb"] Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.298872 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb"] Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328572 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85126e1b-2b92-4e55-b847-d55f8b1b387e-serving-cert\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328668 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6452d\" (UniqueName: \"kubernetes.io/projected/85126e1b-2b92-4e55-b847-d55f8b1b387e-kube-api-access-6452d\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328724 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-config\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328743 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-client-ca\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328776 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qnmc\" (UniqueName: \"kubernetes.io/projected/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-kube-api-access-2qnmc\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328787 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328796 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.328805 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.330439 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-client-ca\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.330543 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-config\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.333309 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85126e1b-2b92-4e55-b847-d55f8b1b387e-serving-cert\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.345408 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6452d\" (UniqueName: \"kubernetes.io/projected/85126e1b-2b92-4e55-b847-d55f8b1b387e-kube-api-access-6452d\") pod \"route-controller-manager-887dfdc8b-vh62w\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.561623 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.564496 4778 patch_prober.go:28] interesting pod/controller-manager-7b56f5b6c6-7q5xm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.564545 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.579243 4778 patch_prober.go:28] interesting pod/route-controller-manager-5f6cfcbfb9-jcsqb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 13:13:26 crc kubenswrapper[4778]: I0312 13:13:26.579294 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5f6cfcbfb9-jcsqb" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 13:13:27 crc kubenswrapper[4778]: I0312 13:13:27.332902 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kc7s7" Mar 12 13:13:28 crc kubenswrapper[4778]: I0312 13:13:28.261851 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fa0e405-9e9d-49fc-b2aa-17ca5c529a74" path="/var/lib/kubelet/pods/1fa0e405-9e9d-49fc-b2aa-17ca5c529a74/volumes" Mar 12 13:13:28 crc kubenswrapper[4778]: I0312 13:13:28.558236 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:13:28 crc kubenswrapper[4778]: I0312 13:13:28.558365 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.212805 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.213936 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.215992 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.216805 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.232270 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.290887 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3834d547-946f-4567-b68b-5305589c5573-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.290971 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3834d547-946f-4567-b68b-5305589c5573-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.392321 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3834d547-946f-4567-b68b-5305589c5573-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.392401 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3834d547-946f-4567-b68b-5305589c5573-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.392581 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3834d547-946f-4567-b68b-5305589c5573-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.428278 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3834d547-946f-4567-b68b-5305589c5573-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:30 crc kubenswrapper[4778]: I0312 13:13:30.538750 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.853414 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.898595 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-d44c8b88d-jx574"] Mar 12 13:13:33 crc kubenswrapper[4778]: E0312 13:13:33.899151 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" containerName="controller-manager" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.899270 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" containerName="controller-manager" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.899474 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" containerName="controller-manager" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.899954 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.912345 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d44c8b88d-jx574"] Mar 12 13:13:33 crc kubenswrapper[4778]: E0312 13:13:33.928170 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 12 13:13:33 crc kubenswrapper[4778]: E0312 13:13:33.928540 4778 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 12 13:13:33 crc kubenswrapper[4778]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 12 13:13:33 crc kubenswrapper[4778]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b6h7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29555352-q7fvr_openshift-infra(9f210efd-2ac0-4b67-89c5-fcd9f52f6e01): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 12 13:13:33 crc kubenswrapper[4778]: > logger="UnhandledError" Mar 12 13:13:33 crc kubenswrapper[4778]: E0312 13:13:33.930003 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" podUID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.943765 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-config\") pod \"b330900c-c52a-4e88-a2d2-38e34f837004\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.943870 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b330900c-c52a-4e88-a2d2-38e34f837004-serving-cert\") pod \"b330900c-c52a-4e88-a2d2-38e34f837004\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.943949 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-proxy-ca-bundles\") pod \"b330900c-c52a-4e88-a2d2-38e34f837004\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.943967 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lwnl\" (UniqueName: \"kubernetes.io/projected/b330900c-c52a-4e88-a2d2-38e34f837004-kube-api-access-4lwnl\") pod \"b330900c-c52a-4e88-a2d2-38e34f837004\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.943984 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-client-ca\") pod \"b330900c-c52a-4e88-a2d2-38e34f837004\" (UID: \"b330900c-c52a-4e88-a2d2-38e34f837004\") " Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.944134 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwdxl\" (UniqueName: \"kubernetes.io/projected/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-kube-api-access-zwdxl\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.944211 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-serving-cert\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.944233 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-config\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.944261 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-client-ca\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.944303 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-proxy-ca-bundles\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.946142 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b330900c-c52a-4e88-a2d2-38e34f837004" (UID: "b330900c-c52a-4e88-a2d2-38e34f837004"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.948768 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-client-ca" (OuterVolumeSpecName: "client-ca") pod "b330900c-c52a-4e88-a2d2-38e34f837004" (UID: "b330900c-c52a-4e88-a2d2-38e34f837004"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.951505 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-config" (OuterVolumeSpecName: "config") pod "b330900c-c52a-4e88-a2d2-38e34f837004" (UID: "b330900c-c52a-4e88-a2d2-38e34f837004"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.977311 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w"] Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.978899 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b330900c-c52a-4e88-a2d2-38e34f837004-kube-api-access-4lwnl" (OuterVolumeSpecName: "kube-api-access-4lwnl") pod "b330900c-c52a-4e88-a2d2-38e34f837004" (UID: "b330900c-c52a-4e88-a2d2-38e34f837004"). InnerVolumeSpecName "kube-api-access-4lwnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:13:33 crc kubenswrapper[4778]: I0312 13:13:33.980457 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b330900c-c52a-4e88-a2d2-38e34f837004-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b330900c-c52a-4e88-a2d2-38e34f837004" (UID: "b330900c-c52a-4e88-a2d2-38e34f837004"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.045712 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-serving-cert\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046186 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-config\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046249 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-client-ca\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046281 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-proxy-ca-bundles\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046314 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwdxl\" (UniqueName: \"kubernetes.io/projected/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-kube-api-access-zwdxl\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046357 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046367 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lwnl\" (UniqueName: \"kubernetes.io/projected/b330900c-c52a-4e88-a2d2-38e34f837004-kube-api-access-4lwnl\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046379 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046387 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b330900c-c52a-4e88-a2d2-38e34f837004-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.046397 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b330900c-c52a-4e88-a2d2-38e34f837004-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.048075 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-config\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.048258 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-client-ca\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.049117 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-proxy-ca-bundles\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.049644 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-serving-cert\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.076462 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwdxl\" (UniqueName: \"kubernetes.io/projected/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-kube-api-access-zwdxl\") pod \"controller-manager-d44c8b88d-jx574\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.216169 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.315786 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" event={"ID":"b330900c-c52a-4e88-a2d2-38e34f837004","Type":"ContainerDied","Data":"17004ca7182405d5ca539723da4c058b3a73eeda64b060c4ba9deebe55691d77"} Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.315862 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm" Mar 12 13:13:34 crc kubenswrapper[4778]: E0312 13:13:34.317663 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" podUID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.351618 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm"] Mar 12 13:13:34 crc kubenswrapper[4778]: I0312 13:13:34.354915 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7b56f5b6c6-7q5xm"] Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.408247 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.408941 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.422411 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.464909 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a868c6a4-19ec-46be-a0af-be25b1049ff3-kube-api-access\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.464967 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.465036 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-var-lock\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.565994 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-var-lock\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.566275 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a868c6a4-19ec-46be-a0af-be25b1049ff3-kube-api-access\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.566116 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-var-lock\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.566315 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.566375 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.600039 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a868c6a4-19ec-46be-a0af-be25b1049ff3-kube-api-access\") pod \"installer-9-crc\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:35 crc kubenswrapper[4778]: I0312 13:13:35.741699 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:13:36 crc kubenswrapper[4778]: I0312 13:13:36.262148 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b330900c-c52a-4e88-a2d2-38e34f837004" path="/var/lib/kubelet/pods/b330900c-c52a-4e88-a2d2-38e34f837004/volumes" Mar 12 13:13:41 crc kubenswrapper[4778]: I0312 13:13:41.815330 4778 scope.go:117] "RemoveContainer" containerID="af0097b4c8ffcf21c4d0f3d542c30c13b992c9bb5a36537354858b4fa3539991" Mar 12 13:13:41 crc kubenswrapper[4778]: E0312 13:13:41.937010 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 12 13:13:41 crc kubenswrapper[4778]: E0312 13:13:41.937471 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cg6lb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-76s88_openshift-marketplace(34ecd758-517c-455a-939a-7eb6d3546854): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:41 crc kubenswrapper[4778]: E0312 13:13:41.938666 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-76s88" podUID="34ecd758-517c-455a-939a-7eb6d3546854" Mar 12 13:13:42 crc kubenswrapper[4778]: I0312 13:13:42.254703 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-rz9vw"] Mar 12 13:13:43 crc kubenswrapper[4778]: E0312 13:13:43.306006 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-76s88" podUID="34ecd758-517c-455a-939a-7eb6d3546854" Mar 12 13:13:43 crc kubenswrapper[4778]: E0312 13:13:43.398776 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 12 13:13:43 crc kubenswrapper[4778]: E0312 13:13:43.398950 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kch8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8xksl_openshift-marketplace(de4557b4-7957-47a0-8c42-845be1fa0f32): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:43 crc kubenswrapper[4778]: E0312 13:13:43.400381 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8xksl" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" Mar 12 13:13:45 crc kubenswrapper[4778]: E0312 13:13:45.064447 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 12 13:13:45 crc kubenswrapper[4778]: E0312 13:13:45.064665 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dxz76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-l8n9b_openshift-marketplace(c27afe2a-3402-49f9-b985-45fe67e40d22): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:45 crc kubenswrapper[4778]: E0312 13:13:45.065856 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-l8n9b" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" Mar 12 13:13:45 crc kubenswrapper[4778]: E0312 13:13:45.273955 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 12 13:13:45 crc kubenswrapper[4778]: E0312 13:13:45.274204 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mpfz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5s5vs_openshift-marketplace(f438f2a3-60c0-4554-a49b-030545f8139c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:45 crc kubenswrapper[4778]: E0312 13:13:45.275848 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5s5vs" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.166654 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8xksl" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.172643 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5s5vs" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.173128 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-l8n9b" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" Mar 12 13:13:46 crc kubenswrapper[4778]: W0312 13:13:46.176325 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-d04218a9594ee2e8723cf07ab5927db677dc447d864eaf55e7287a809c797b89 WatchSource:0}: Error finding container d04218a9594ee2e8723cf07ab5927db677dc447d864eaf55e7287a809c797b89: Status 404 returned error can't find the container with id d04218a9594ee2e8723cf07ab5927db677dc447d864eaf55e7287a809c797b89 Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.282708 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.282916 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dq7jg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-sjk9p_openshift-marketplace(3b3fb69e-dd4f-4787-a207-4fe25106f9e7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.284677 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-sjk9p" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.389602 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b57645a462ca0d3c7886fbff7aa76102f8895a353f74fc78b7de9ff180fc4427"} Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.397904 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" event={"ID":"0b59b25a-3acc-4d06-b91d-575f45463520","Type":"ContainerStarted","Data":"36557cc8ad045b218ec496388cee2d6541a247487ec1e3da23520d22518f4d28"} Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.400642 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e7b95dbba219e1b176f09463e4a0afd5bb966d4ae2a38107f3e099cd2cd3b809"} Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.403577 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d04218a9594ee2e8723cf07ab5927db677dc447d864eaf55e7287a809c797b89"} Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.407026 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-sjk9p" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.483667 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 13:13:46 crc kubenswrapper[4778]: W0312 13:13:46.500394 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda868c6a4_19ec_46be_a0af_be25b1049ff3.slice/crio-67b06efe996403c5470e41a5f9a62a78fe522b551d7ec62d8302163676162a07 WatchSource:0}: Error finding container 67b06efe996403c5470e41a5f9a62a78fe522b551d7ec62d8302163676162a07: Status 404 returned error can't find the container with id 67b06efe996403c5470e41a5f9a62a78fe522b551d7ec62d8302163676162a07 Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.511434 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.511631 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6mx5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qx9d8_openshift-marketplace(651601bd-18fe-4ca1-9c61-481ca568d022): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.514002 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qx9d8" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.544743 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d44c8b88d-jx574"] Mar 12 13:13:46 crc kubenswrapper[4778]: W0312 13:13:46.561487 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fdd5690_0e80_4317_9e3a_8478f09ea1a8.slice/crio-dd0b16a4e92ddcc0e1151ae83fdba0245e8931b7997e602261eaa93e0a982440 WatchSource:0}: Error finding container dd0b16a4e92ddcc0e1151ae83fdba0245e8931b7997e602261eaa93e0a982440: Status 404 returned error can't find the container with id dd0b16a4e92ddcc0e1151ae83fdba0245e8931b7997e602261eaa93e0a982440 Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.634531 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 13:13:46 crc kubenswrapper[4778]: I0312 13:13:46.643112 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w"] Mar 12 13:13:46 crc kubenswrapper[4778]: W0312 13:13:46.651970 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3834d547_946f_4567_b68b_5305589c5573.slice/crio-a78d297f108842c664691a098601755963d8c21b84c611f8a3c408f8ad8e233d WatchSource:0}: Error finding container a78d297f108842c664691a098601755963d8c21b84c611f8a3c408f8ad8e233d: Status 404 returned error can't find the container with id a78d297f108842c664691a098601755963d8c21b84c611f8a3c408f8ad8e233d Mar 12 13:13:46 crc kubenswrapper[4778]: W0312 13:13:46.655683 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85126e1b_2b92_4e55_b847_d55f8b1b387e.slice/crio-d09bd42fb521e8e0212e9f4502cde76e14304082c90f8d8f8186fe008442beee WatchSource:0}: Error finding container d09bd42fb521e8e0212e9f4502cde76e14304082c90f8d8f8186fe008442beee: Status 404 returned error can't find the container with id d09bd42fb521e8e0212e9f4502cde76e14304082c90f8d8f8186fe008442beee Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.763983 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.764150 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzdp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-khr6h_openshift-marketplace(1d185732-cd6b-44c6-b4db-ee9ade00c683): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.765558 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-khr6h" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.874231 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.874389 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gjm7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rtjz5_openshift-marketplace(b9bef112-9bef-4ce2-abd8-054b4d671658): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 13:13:46 crc kubenswrapper[4778]: E0312 13:13:46.875772 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rtjz5" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.410866 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3834d547-946f-4567-b68b-5305589c5573","Type":"ContainerStarted","Data":"85369356d1ab8bd11065ca773167d0e1195b4fa7747a6ade645e0d61aacbd264"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.411224 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3834d547-946f-4567-b68b-5305589c5573","Type":"ContainerStarted","Data":"a78d297f108842c664691a098601755963d8c21b84c611f8a3c408f8ad8e233d"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.412117 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c86f994ca77821cf425217729aab1099a7f4898a6c96f51e520a438bc332ecdd"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.412151 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.413753 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" event={"ID":"0b59b25a-3acc-4d06-b91d-575f45463520","Type":"ContainerStarted","Data":"ed9b865a7a64df4df0449be2c8d2e0dba2b2e0d5df9dac000ddb65d397b37f8b"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.413782 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rz9vw" event={"ID":"0b59b25a-3acc-4d06-b91d-575f45463520","Type":"ContainerStarted","Data":"c9663cee45995acefceb1213bb8b5f6c578e1c342fb22d2b298df0e1a4b6f7ea"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.414721 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a868c6a4-19ec-46be-a0af-be25b1049ff3","Type":"ContainerStarted","Data":"35dc89f42df73eafd54f7518d380b5b4f6934732de9c6dd0209b64b9345aa66c"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.414751 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a868c6a4-19ec-46be-a0af-be25b1049ff3","Type":"ContainerStarted","Data":"67b06efe996403c5470e41a5f9a62a78fe522b551d7ec62d8302163676162a07"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.415849 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" event={"ID":"85126e1b-2b92-4e55-b847-d55f8b1b387e","Type":"ContainerStarted","Data":"787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.415879 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" event={"ID":"85126e1b-2b92-4e55-b847-d55f8b1b387e","Type":"ContainerStarted","Data":"d09bd42fb521e8e0212e9f4502cde76e14304082c90f8d8f8186fe008442beee"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.415881 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" podUID="85126e1b-2b92-4e55-b847-d55f8b1b387e" containerName="route-controller-manager" containerID="cri-o://787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b" gracePeriod=30 Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.415994 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.418063 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c2182121d64bdd93ec58bbd5463eeb9f34a0319fc75e80286d85a3e7087bbd62"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.420415 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.425890 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9b1b3fd04d6e452de07ce84a8d13b430669a8c1b705e59aa538d8f1a53cc86b9"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.427466 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" event={"ID":"0fdd5690-0e80-4317-9e3a-8478f09ea1a8","Type":"ContainerStarted","Data":"de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.427493 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" event={"ID":"0fdd5690-0e80-4317-9e3a-8478f09ea1a8","Type":"ContainerStarted","Data":"dd0b16a4e92ddcc0e1151ae83fdba0245e8931b7997e602261eaa93e0a982440"} Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.427711 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:47 crc kubenswrapper[4778]: E0312 13:13:47.428842 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-khr6h" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" Mar 12 13:13:47 crc kubenswrapper[4778]: E0312 13:13:47.428955 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rtjz5" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" Mar 12 13:13:47 crc kubenswrapper[4778]: E0312 13:13:47.431838 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qx9d8" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.432844 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.438856 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=17.438835269 podStartE2EDuration="17.438835269s" podCreationTimestamp="2026-03-12 13:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:13:47.433692824 +0000 UTC m=+245.882388220" watchObservedRunningTime="2026-03-12 13:13:47.438835269 +0000 UTC m=+245.887530665" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.468268 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=12.468249518 podStartE2EDuration="12.468249518s" podCreationTimestamp="2026-03-12 13:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:13:47.466735215 +0000 UTC m=+245.915430621" watchObservedRunningTime="2026-03-12 13:13:47.468249518 +0000 UTC m=+245.916944924" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.470028 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-rz9vw" podStartSLOduration=196.470017948 podStartE2EDuration="3m16.470017948s" podCreationTimestamp="2026-03-12 13:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:13:47.45306149 +0000 UTC m=+245.901756886" watchObservedRunningTime="2026-03-12 13:13:47.470017948 +0000 UTC m=+245.918713344" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.569787 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" podStartSLOduration=34.569762129 podStartE2EDuration="34.569762129s" podCreationTimestamp="2026-03-12 13:13:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:13:47.56943292 +0000 UTC m=+246.018128316" watchObservedRunningTime="2026-03-12 13:13:47.569762129 +0000 UTC m=+246.018457525" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.614374 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" podStartSLOduration=14.614335035 podStartE2EDuration="14.614335035s" podCreationTimestamp="2026-03-12 13:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:13:47.592892001 +0000 UTC m=+246.041587397" watchObservedRunningTime="2026-03-12 13:13:47.614335035 +0000 UTC m=+246.063030431" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.853328 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.883830 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5"] Mar 12 13:13:47 crc kubenswrapper[4778]: E0312 13:13:47.884075 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85126e1b-2b92-4e55-b847-d55f8b1b387e" containerName="route-controller-manager" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.884087 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="85126e1b-2b92-4e55-b847-d55f8b1b387e" containerName="route-controller-manager" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.884183 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="85126e1b-2b92-4e55-b847-d55f8b1b387e" containerName="route-controller-manager" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.884793 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.895733 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5"] Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950297 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-client-ca\") pod \"85126e1b-2b92-4e55-b847-d55f8b1b387e\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950346 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6452d\" (UniqueName: \"kubernetes.io/projected/85126e1b-2b92-4e55-b847-d55f8b1b387e-kube-api-access-6452d\") pod \"85126e1b-2b92-4e55-b847-d55f8b1b387e\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950416 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-config\") pod \"85126e1b-2b92-4e55-b847-d55f8b1b387e\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950433 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85126e1b-2b92-4e55-b847-d55f8b1b387e-serving-cert\") pod \"85126e1b-2b92-4e55-b847-d55f8b1b387e\" (UID: \"85126e1b-2b92-4e55-b847-d55f8b1b387e\") " Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950596 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/badeb3df-9c56-4aa2-af6f-aba14c213fcc-serving-cert\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950723 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-config\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950752 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcsqr\" (UniqueName: \"kubernetes.io/projected/badeb3df-9c56-4aa2-af6f-aba14c213fcc-kube-api-access-tcsqr\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950772 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-client-ca\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.950990 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-client-ca" (OuterVolumeSpecName: "client-ca") pod "85126e1b-2b92-4e55-b847-d55f8b1b387e" (UID: "85126e1b-2b92-4e55-b847-d55f8b1b387e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.951341 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-config" (OuterVolumeSpecName: "config") pod "85126e1b-2b92-4e55-b847-d55f8b1b387e" (UID: "85126e1b-2b92-4e55-b847-d55f8b1b387e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.955534 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85126e1b-2b92-4e55-b847-d55f8b1b387e-kube-api-access-6452d" (OuterVolumeSpecName: "kube-api-access-6452d") pod "85126e1b-2b92-4e55-b847-d55f8b1b387e" (UID: "85126e1b-2b92-4e55-b847-d55f8b1b387e"). InnerVolumeSpecName "kube-api-access-6452d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:13:47 crc kubenswrapper[4778]: I0312 13:13:47.957011 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85126e1b-2b92-4e55-b847-d55f8b1b387e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "85126e1b-2b92-4e55-b847-d55f8b1b387e" (UID: "85126e1b-2b92-4e55-b847-d55f8b1b387e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.051791 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcsqr\" (UniqueName: \"kubernetes.io/projected/badeb3df-9c56-4aa2-af6f-aba14c213fcc-kube-api-access-tcsqr\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.051860 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-client-ca\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.051918 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/badeb3df-9c56-4aa2-af6f-aba14c213fcc-serving-cert\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.051970 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-config\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.052013 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.052025 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6452d\" (UniqueName: \"kubernetes.io/projected/85126e1b-2b92-4e55-b847-d55f8b1b387e-kube-api-access-6452d\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.052039 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85126e1b-2b92-4e55-b847-d55f8b1b387e-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.052050 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85126e1b-2b92-4e55-b847-d55f8b1b387e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.053373 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-config\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.054678 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-client-ca\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.057823 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/badeb3df-9c56-4aa2-af6f-aba14c213fcc-serving-cert\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.075108 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcsqr\" (UniqueName: \"kubernetes.io/projected/badeb3df-9c56-4aa2-af6f-aba14c213fcc-kube-api-access-tcsqr\") pod \"route-controller-manager-6f4dd5cc6-ppsx5\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.207232 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.435075 4778 generic.go:334] "Generic (PLEG): container finished" podID="3834d547-946f-4567-b68b-5305589c5573" containerID="85369356d1ab8bd11065ca773167d0e1195b4fa7747a6ade645e0d61aacbd264" exitCode=0 Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.435148 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3834d547-946f-4567-b68b-5305589c5573","Type":"ContainerDied","Data":"85369356d1ab8bd11065ca773167d0e1195b4fa7747a6ade645e0d61aacbd264"} Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.437481 4778 generic.go:334] "Generic (PLEG): container finished" podID="85126e1b-2b92-4e55-b847-d55f8b1b387e" containerID="787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b" exitCode=0 Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.437572 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.437632 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" event={"ID":"85126e1b-2b92-4e55-b847-d55f8b1b387e","Type":"ContainerDied","Data":"787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b"} Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.437669 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w" event={"ID":"85126e1b-2b92-4e55-b847-d55f8b1b387e","Type":"ContainerDied","Data":"d09bd42fb521e8e0212e9f4502cde76e14304082c90f8d8f8186fe008442beee"} Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.437690 4778 scope.go:117] "RemoveContainer" containerID="787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.465719 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w"] Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.469711 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-887dfdc8b-vh62w"] Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.469945 4778 scope.go:117] "RemoveContainer" containerID="787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b" Mar 12 13:13:48 crc kubenswrapper[4778]: E0312 13:13:48.470449 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b\": container with ID starting with 787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b not found: ID does not exist" containerID="787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.470475 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b"} err="failed to get container status \"787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b\": rpc error: code = NotFound desc = could not find container \"787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b\": container with ID starting with 787bc707ec53d701471888ee4a1be3c5289a94da6f7f14cfd6a0ac8a2a7cbd1b not found: ID does not exist" Mar 12 13:13:48 crc kubenswrapper[4778]: I0312 13:13:48.666504 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5"] Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.180614 4778 csr.go:261] certificate signing request csr-twzvk is approved, waiting to be issued Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.189952 4778 csr.go:257] certificate signing request csr-twzvk is issued Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.447556 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" event={"ID":"badeb3df-9c56-4aa2-af6f-aba14c213fcc","Type":"ContainerStarted","Data":"6edfc1174eae36c4699c23b09d94a6801a70e404d52b1e50d4350988d1f6d371"} Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.447610 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" event={"ID":"badeb3df-9c56-4aa2-af6f-aba14c213fcc","Type":"ContainerStarted","Data":"ad682c14c40bcfdf45bab0f4aae014cbaeacf6b49ffe857ff368861fb7bbc412"} Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.447829 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.450292 4778 generic.go:334] "Generic (PLEG): container finished" podID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" containerID="d6a4e00222817c0335bb85eb95073d869a129a695fed4bc12743392acf13e251" exitCode=0 Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.450377 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" event={"ID":"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01","Type":"ContainerDied","Data":"d6a4e00222817c0335bb85eb95073d869a129a695fed4bc12743392acf13e251"} Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.455972 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.487815 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" podStartSLOduration=16.487793128 podStartE2EDuration="16.487793128s" podCreationTimestamp="2026-03-12 13:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:13:49.468025051 +0000 UTC m=+247.916720447" watchObservedRunningTime="2026-03-12 13:13:49.487793128 +0000 UTC m=+247.936488524" Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.765983 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.905602 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3834d547-946f-4567-b68b-5305589c5573-kube-api-access\") pod \"3834d547-946f-4567-b68b-5305589c5573\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.905705 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3834d547-946f-4567-b68b-5305589c5573-kubelet-dir\") pod \"3834d547-946f-4567-b68b-5305589c5573\" (UID: \"3834d547-946f-4567-b68b-5305589c5573\") " Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.905806 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3834d547-946f-4567-b68b-5305589c5573-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3834d547-946f-4567-b68b-5305589c5573" (UID: "3834d547-946f-4567-b68b-5305589c5573"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.906153 4778 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3834d547-946f-4567-b68b-5305589c5573-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:49 crc kubenswrapper[4778]: I0312 13:13:49.912489 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3834d547-946f-4567-b68b-5305589c5573-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3834d547-946f-4567-b68b-5305589c5573" (UID: "3834d547-946f-4567-b68b-5305589c5573"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.007842 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3834d547-946f-4567-b68b-5305589c5573-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.191146 4778 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-20 13:35:55.351213448 +0000 UTC Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.191617 4778 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6072h22m5.159692805s for next certificate rotation Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.265313 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85126e1b-2b92-4e55-b847-d55f8b1b387e" path="/var/lib/kubelet/pods/85126e1b-2b92-4e55-b847-d55f8b1b387e/volumes" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.462868 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.463371 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3834d547-946f-4567-b68b-5305589c5573","Type":"ContainerDied","Data":"a78d297f108842c664691a098601755963d8c21b84c611f8a3c408f8ad8e233d"} Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.463431 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a78d297f108842c664691a098601755963d8c21b84c611f8a3c408f8ad8e233d" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.717822 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.820481 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6h7h\" (UniqueName: \"kubernetes.io/projected/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01-kube-api-access-b6h7h\") pod \"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01\" (UID: \"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01\") " Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.826292 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01-kube-api-access-b6h7h" (OuterVolumeSpecName: "kube-api-access-b6h7h") pod "9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" (UID: "9f210efd-2ac0-4b67-89c5-fcd9f52f6e01"). InnerVolumeSpecName "kube-api-access-b6h7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:13:50 crc kubenswrapper[4778]: I0312 13:13:50.921884 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6h7h\" (UniqueName: \"kubernetes.io/projected/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01-kube-api-access-b6h7h\") on node \"crc\" DevicePath \"\"" Mar 12 13:13:51 crc kubenswrapper[4778]: I0312 13:13:51.192440 4778 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-13 10:03:21.382508358 +0000 UTC Mar 12 13:13:51 crc kubenswrapper[4778]: I0312 13:13:51.192480 4778 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6620h49m30.190030163s for next certificate rotation Mar 12 13:13:51 crc kubenswrapper[4778]: I0312 13:13:51.469873 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" event={"ID":"9f210efd-2ac0-4b67-89c5-fcd9f52f6e01","Type":"ContainerDied","Data":"0a2c8918cbacef5d63ed30076a63c59219bb878177978f4909e3ed43cb24db19"} Mar 12 13:13:51 crc kubenswrapper[4778]: I0312 13:13:51.469945 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a2c8918cbacef5d63ed30076a63c59219bb878177978f4909e3ed43cb24db19" Mar 12 13:13:51 crc kubenswrapper[4778]: I0312 13:13:51.469899 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555352-q7fvr" Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.508385 4778 generic.go:334] "Generic (PLEG): container finished" podID="34ecd758-517c-455a-939a-7eb6d3546854" containerID="e6857324d1a49d08837ab795e083cf8ed33ad61f45f62f385bd7494ef38b2514" exitCode=0 Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.508428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerDied","Data":"e6857324d1a49d08837ab795e083cf8ed33ad61f45f62f385bd7494ef38b2514"} Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.558226 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.558284 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.558329 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.558965 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:13:58 crc kubenswrapper[4778]: I0312 13:13:58.559024 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce" gracePeriod=600 Mar 12 13:13:59 crc kubenswrapper[4778]: I0312 13:13:59.515322 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce" exitCode=0 Mar 12 13:13:59 crc kubenswrapper[4778]: I0312 13:13:59.515469 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce"} Mar 12 13:13:59 crc kubenswrapper[4778]: I0312 13:13:59.515995 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"dcabd48eda797c052967d086d455193bf30a1f05151385a52352d733c58148f7"} Mar 12 13:13:59 crc kubenswrapper[4778]: I0312 13:13:59.526584 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerStarted","Data":"ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354"} Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.139629 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555354-n6zvc"] Mar 12 13:14:00 crc kubenswrapper[4778]: E0312 13:14:00.141419 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" containerName="oc" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.141455 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" containerName="oc" Mar 12 13:14:00 crc kubenswrapper[4778]: E0312 13:14:00.141478 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3834d547-946f-4567-b68b-5305589c5573" containerName="pruner" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.141491 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3834d547-946f-4567-b68b-5305589c5573" containerName="pruner" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.141696 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" containerName="oc" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.141716 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3834d547-946f-4567-b68b-5305589c5573" containerName="pruner" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.142284 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.145456 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555354-n6zvc"] Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.180917 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.181370 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.181542 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.181984 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbzs\" (UniqueName: \"kubernetes.io/projected/f91620d9-a95e-4e74-ab13-531d5e040b50-kube-api-access-vxbzs\") pod \"auto-csr-approver-29555354-n6zvc\" (UID: \"f91620d9-a95e-4e74-ab13-531d5e040b50\") " pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.283151 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbzs\" (UniqueName: \"kubernetes.io/projected/f91620d9-a95e-4e74-ab13-531d5e040b50-kube-api-access-vxbzs\") pod \"auto-csr-approver-29555354-n6zvc\" (UID: \"f91620d9-a95e-4e74-ab13-531d5e040b50\") " pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.302803 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbzs\" (UniqueName: \"kubernetes.io/projected/f91620d9-a95e-4e74-ab13-531d5e040b50-kube-api-access-vxbzs\") pod \"auto-csr-approver-29555354-n6zvc\" (UID: \"f91620d9-a95e-4e74-ab13-531d5e040b50\") " pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:00 crc kubenswrapper[4778]: I0312 13:14:00.502091 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:01 crc kubenswrapper[4778]: I0312 13:14:01.291521 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555354-n6zvc"] Mar 12 13:14:01 crc kubenswrapper[4778]: W0312 13:14:01.294094 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf91620d9_a95e_4e74_ab13_531d5e040b50.slice/crio-48edef7d58f43f0f3d3ff5833c8716761e6649acb2b5f5bb2b27f21563634198 WatchSource:0}: Error finding container 48edef7d58f43f0f3d3ff5833c8716761e6649acb2b5f5bb2b27f21563634198: Status 404 returned error can't find the container with id 48edef7d58f43f0f3d3ff5833c8716761e6649acb2b5f5bb2b27f21563634198 Mar 12 13:14:01 crc kubenswrapper[4778]: I0312 13:14:01.546006 4778 generic.go:334] "Generic (PLEG): container finished" podID="f438f2a3-60c0-4554-a49b-030545f8139c" containerID="ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354" exitCode=0 Mar 12 13:14:01 crc kubenswrapper[4778]: I0312 13:14:01.546082 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerDied","Data":"ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354"} Mar 12 13:14:01 crc kubenswrapper[4778]: I0312 13:14:01.548382 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" event={"ID":"f91620d9-a95e-4e74-ab13-531d5e040b50","Type":"ContainerStarted","Data":"48edef7d58f43f0f3d3ff5833c8716761e6649acb2b5f5bb2b27f21563634198"} Mar 12 13:14:08 crc kubenswrapper[4778]: I0312 13:14:08.593996 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerStarted","Data":"1e77f31cb8ac97bbace99ce9835f811074e891b28dabf061e7039bfab7607d57"} Mar 12 13:14:08 crc kubenswrapper[4778]: I0312 13:14:08.616817 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-76s88" podStartSLOduration=2.043036075 podStartE2EDuration="1m11.616781986s" podCreationTimestamp="2026-03-12 13:12:57 +0000 UTC" firstStartedPulling="2026-03-12 13:12:58.043338486 +0000 UTC m=+196.492033882" lastFinishedPulling="2026-03-12 13:14:07.617084387 +0000 UTC m=+266.065779793" observedRunningTime="2026-03-12 13:14:08.612690754 +0000 UTC m=+267.061386150" watchObservedRunningTime="2026-03-12 13:14:08.616781986 +0000 UTC m=+267.065477382" Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.611735 4778 generic.go:334] "Generic (PLEG): container finished" podID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerID="718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.611861 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xksl" event={"ID":"de4557b4-7957-47a0-8c42-845be1fa0f32","Type":"ContainerDied","Data":"718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.620810 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerStarted","Data":"06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.625282 4778 generic.go:334] "Generic (PLEG): container finished" podID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerID="fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.625279 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjk9p" event={"ID":"3b3fb69e-dd4f-4787-a207-4fe25106f9e7","Type":"ContainerDied","Data":"fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.627728 4778 generic.go:334] "Generic (PLEG): container finished" podID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerID="84fe3c954d7e0d1d6303467d2621bf3b31d896882603252deb19491a2fa354ed" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.627785 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khr6h" event={"ID":"1d185732-cd6b-44c6-b4db-ee9ade00c683","Type":"ContainerDied","Data":"84fe3c954d7e0d1d6303467d2621bf3b31d896882603252deb19491a2fa354ed"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.636152 4778 generic.go:334] "Generic (PLEG): container finished" podID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerID="44212f253b9d8de159bf039fe64dd134b5f7beb71943da6aab7d4efc080466b3" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.636582 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtjz5" event={"ID":"b9bef112-9bef-4ce2-abd8-054b4d671658","Type":"ContainerDied","Data":"44212f253b9d8de159bf039fe64dd134b5f7beb71943da6aab7d4efc080466b3"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.641725 4778 generic.go:334] "Generic (PLEG): container finished" podID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerID="517c2af638efb950196e9ef53f4578b28c6c02cc9d241b33a72ede0303af599d" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.641800 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l8n9b" event={"ID":"c27afe2a-3402-49f9-b985-45fe67e40d22","Type":"ContainerDied","Data":"517c2af638efb950196e9ef53f4578b28c6c02cc9d241b33a72ede0303af599d"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.650686 4778 generic.go:334] "Generic (PLEG): container finished" podID="f91620d9-a95e-4e74-ab13-531d5e040b50" containerID="6a586e8ffe815ea410f687edd18208ce93300b26a8a15a7f7b6bd8396c76c788" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.650872 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" event={"ID":"f91620d9-a95e-4e74-ab13-531d5e040b50","Type":"ContainerDied","Data":"6a586e8ffe815ea410f687edd18208ce93300b26a8a15a7f7b6bd8396c76c788"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.655240 4778 generic.go:334] "Generic (PLEG): container finished" podID="651601bd-18fe-4ca1-9c61-481ca568d022" containerID="777dcb7d13b3c9f17ff760e883a8a2c8d277b3c6622f9924b38301e80f9b85e9" exitCode=0 Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.655287 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qx9d8" event={"ID":"651601bd-18fe-4ca1-9c61-481ca568d022","Type":"ContainerDied","Data":"777dcb7d13b3c9f17ff760e883a8a2c8d277b3c6622f9924b38301e80f9b85e9"} Mar 12 13:14:09 crc kubenswrapper[4778]: I0312 13:14:09.697293 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5s5vs" podStartSLOduration=3.443434404 podStartE2EDuration="1m13.697275483s" podCreationTimestamp="2026-03-12 13:12:56 +0000 UTC" firstStartedPulling="2026-03-12 13:12:58.037477361 +0000 UTC m=+196.486172757" lastFinishedPulling="2026-03-12 13:14:08.29131844 +0000 UTC m=+266.740013836" observedRunningTime="2026-03-12 13:14:09.689799453 +0000 UTC m=+268.138494849" watchObservedRunningTime="2026-03-12 13:14:09.697275483 +0000 UTC m=+268.145970879" Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.032957 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.139035 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxbzs\" (UniqueName: \"kubernetes.io/projected/f91620d9-a95e-4e74-ab13-531d5e040b50-kube-api-access-vxbzs\") pod \"f91620d9-a95e-4e74-ab13-531d5e040b50\" (UID: \"f91620d9-a95e-4e74-ab13-531d5e040b50\") " Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.144259 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91620d9-a95e-4e74-ab13-531d5e040b50-kube-api-access-vxbzs" (OuterVolumeSpecName: "kube-api-access-vxbzs") pod "f91620d9-a95e-4e74-ab13-531d5e040b50" (UID: "f91620d9-a95e-4e74-ab13-531d5e040b50"). InnerVolumeSpecName "kube-api-access-vxbzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.240763 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxbzs\" (UniqueName: \"kubernetes.io/projected/f91620d9-a95e-4e74-ab13-531d5e040b50-kube-api-access-vxbzs\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.668143 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" event={"ID":"f91620d9-a95e-4e74-ab13-531d5e040b50","Type":"ContainerDied","Data":"48edef7d58f43f0f3d3ff5833c8716761e6649acb2b5f5bb2b27f21563634198"} Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.668200 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555354-n6zvc" Mar 12 13:14:11 crc kubenswrapper[4778]: I0312 13:14:11.668211 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48edef7d58f43f0f3d3ff5833c8716761e6649acb2b5f5bb2b27f21563634198" Mar 12 13:14:14 crc kubenswrapper[4778]: I0312 13:14:14.684429 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xksl" event={"ID":"de4557b4-7957-47a0-8c42-845be1fa0f32","Type":"ContainerStarted","Data":"1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592"} Mar 12 13:14:15 crc kubenswrapper[4778]: I0312 13:14:15.757365 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8xksl" podStartSLOduration=3.613980098 podStartE2EDuration="1m20.757346364s" podCreationTimestamp="2026-03-12 13:12:55 +0000 UTC" firstStartedPulling="2026-03-12 13:12:56.999981705 +0000 UTC m=+195.448677101" lastFinishedPulling="2026-03-12 13:14:14.143347981 +0000 UTC m=+272.592043367" observedRunningTime="2026-03-12 13:14:15.75252867 +0000 UTC m=+274.201224066" watchObservedRunningTime="2026-03-12 13:14:15.757346364 +0000 UTC m=+274.206041760" Mar 12 13:14:16 crc kubenswrapper[4778]: I0312 13:14:16.290718 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 13:14:16 crc kubenswrapper[4778]: I0312 13:14:16.811321 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kw4v"] Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.008409 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.008459 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.481048 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.481086 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.892943 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.893021 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.937911 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:14:17 crc kubenswrapper[4778]: I0312 13:14:17.943305 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:14:19 crc kubenswrapper[4778]: I0312 13:14:19.119830 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-76s88"] Mar 12 13:14:19 crc kubenswrapper[4778]: I0312 13:14:19.755325 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-76s88" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="registry-server" containerID="cri-o://1e77f31cb8ac97bbace99ce9835f811074e891b28dabf061e7039bfab7607d57" gracePeriod=2 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:22.788593 4778 generic.go:334] "Generic (PLEG): container finished" podID="34ecd758-517c-455a-939a-7eb6d3546854" containerID="1e77f31cb8ac97bbace99ce9835f811074e891b28dabf061e7039bfab7607d57" exitCode=0 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:22.788659 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerDied","Data":"1e77f31cb8ac97bbace99ce9835f811074e891b28dabf061e7039bfab7607d57"} Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.440928 4778 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.441526 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91620d9-a95e-4e74-ab13-531d5e040b50" containerName="oc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.441539 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91620d9-a95e-4e74-ab13-531d5e040b50" containerName="oc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.441645 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f91620d9-a95e-4e74-ab13-531d5e040b50" containerName="oc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.441959 4778 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442052 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442303 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1" gracePeriod=15 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442374 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51" gracePeriod=15 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442383 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0" gracePeriod=15 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442383 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c" gracePeriod=15 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442463 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a" gracePeriod=15 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442776 4778 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.442925 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442938 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.442947 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442954 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.442964 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442971 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.442979 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.442986 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.442997 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443003 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.443016 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443024 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.443033 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443040 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.443047 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443053 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.443064 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443069 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443178 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443216 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443225 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443236 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443248 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443256 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443265 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443273 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: E0312 13:14:24.443420 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443432 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.443548 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.468023 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.550139 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.550260 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.550456 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.550684 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.550868 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.551108 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.551174 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.551318 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653052 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653109 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653130 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653167 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653222 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653243 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653241 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653280 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653159 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653319 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653365 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653403 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653427 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653446 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653482 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.653542 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.765522 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.803960 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.805056 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.805673 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a" exitCode=2 Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.815728 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76s88" event={"ID":"34ecd758-517c-455a-939a-7eb6d3546854","Type":"ContainerDied","Data":"cf68cb478854e264cd59c9ad8e9f3e763498e2e2706254a3b88fc3dd9f22fe4f"} Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.815756 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf68cb478854e264cd59c9ad8e9f3e763498e2e2706254a3b88fc3dd9f22fe4f" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.836931 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.837692 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.837980 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.838489 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.957403 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg6lb\" (UniqueName: \"kubernetes.io/projected/34ecd758-517c-455a-939a-7eb6d3546854-kube-api-access-cg6lb\") pod \"34ecd758-517c-455a-939a-7eb6d3546854\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.957687 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-utilities\") pod \"34ecd758-517c-455a-939a-7eb6d3546854\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.957730 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-catalog-content\") pod \"34ecd758-517c-455a-939a-7eb6d3546854\" (UID: \"34ecd758-517c-455a-939a-7eb6d3546854\") " Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.958431 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-utilities" (OuterVolumeSpecName: "utilities") pod "34ecd758-517c-455a-939a-7eb6d3546854" (UID: "34ecd758-517c-455a-939a-7eb6d3546854"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:14:24 crc kubenswrapper[4778]: I0312 13:14:24.962528 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ecd758-517c-455a-939a-7eb6d3546854-kube-api-access-cg6lb" (OuterVolumeSpecName: "kube-api-access-cg6lb") pod "34ecd758-517c-455a-939a-7eb6d3546854" (UID: "34ecd758-517c-455a-939a-7eb6d3546854"). InnerVolumeSpecName "kube-api-access-cg6lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.059310 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg6lb\" (UniqueName: \"kubernetes.io/projected/34ecd758-517c-455a-939a-7eb6d3546854-kube-api-access-cg6lb\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.059341 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.214757 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34ecd758-517c-455a-939a-7eb6d3546854" (UID: "34ecd758-517c-455a-939a-7eb6d3546854"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.262489 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34ecd758-517c-455a-939a-7eb6d3546854-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.634692 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.636267 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.696338 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.697228 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.697611 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.698009 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.698455 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.821973 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.823241 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.823904 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c" exitCode=0 Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.823930 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0" exitCode=0 Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.823938 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51" exitCode=0 Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.823999 4778 scope.go:117] "RemoveContainer" containerID="14c7f2ade3aac502f0534414554216096b45d4f78e81f8ec213064a6205efdbd" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.824002 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76s88" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.824915 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.825816 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.826328 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.826674 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.837862 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.838261 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.838558 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.838799 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.858696 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.859516 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.859982 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.860456 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: I0312 13:14:25.860834 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:25 crc kubenswrapper[4778]: E0312 13:14:25.884626 4778 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.32:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-sjk9p.189c1a4bbda7962d openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-sjk9p,UID:3b3fb69e-dd4f-4787-a207-4fe25106f9e7,APIVersion:v1,ResourceVersion:28313,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 16.256s (16.256s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:14:25.883878957 +0000 UTC m=+284.332574353,LastTimestamp:2026-03-12 13:14:25.883878957 +0000 UTC m=+284.332574353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:14:26 crc kubenswrapper[4778]: W0312 13:14:26.193814 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-dd57d9c1e52f398a835a3a8d9c99f0c1ae6b500c97490f43d9008f6901a63f4c WatchSource:0}: Error finding container dd57d9c1e52f398a835a3a8d9c99f0c1ae6b500c97490f43d9008f6901a63f4c: Status 404 returned error can't find the container with id dd57d9c1e52f398a835a3a8d9c99f0c1ae6b500c97490f43d9008f6901a63f4c Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.493140 4778 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.493600 4778 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.493827 4778 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.494011 4778 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.494374 4778 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: I0312 13:14:26.494428 4778 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.494768 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="200ms" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.709852 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="400ms" Mar 12 13:14:26 crc kubenswrapper[4778]: I0312 13:14:26.829560 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"dd57d9c1e52f398a835a3a8d9c99f0c1ae6b500c97490f43d9008f6901a63f4c"} Mar 12 13:14:26 crc kubenswrapper[4778]: I0312 13:14:26.832765 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.854281 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:26Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:26Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:26Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:26Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:0d4c830b2653f2eeffebd09537afb06afb5ae827adbc03f224ab7269f399c05c\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:d6065909bc521a3f9a85174276fdbceafad02a276449a7dd1952a1f689b0d362\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1735807445},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:185237e125a9d710a58d4b588ea6b75eb361e4e99d979c1acd193de3b5d787f1\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:746054bb64fa0b27b1a696cd5db508bb9ee883a94969e4c1c4b5d35a93da8ef5\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1281521882},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:898c67bf7fc973e99114f3148976a6c21ae0dbe413051415588fa9b995f5b331\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:a641939d2096609a4cf6eec872a1476b7c671bfd81cffc2edeb6e9f13c9deeba\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1231028434},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0e6908b5c2800b56584a3fdf3bc164b76cb945966a49103123dabb61f8e367f2\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:ad31505e97766fe3b9d49abfe33098361de32a828c13e290be714f02a7ee76e0\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1221788890},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.855009 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.855431 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.855760 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.856152 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:26 crc kubenswrapper[4778]: E0312 13:14:26.856252 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:14:27 crc kubenswrapper[4778]: E0312 13:14:27.111269 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="800ms" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.840343 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l8n9b" event={"ID":"c27afe2a-3402-49f9-b985-45fe67e40d22","Type":"ContainerStarted","Data":"3686a4e289950327029466c928723a8314f5dcaa797637ff0db63d9aa4aeb5db"} Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.841457 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.841889 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.843160 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.843199 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qx9d8" event={"ID":"651601bd-18fe-4ca1-9c61-481ca568d022","Type":"ContainerStarted","Data":"13189da41e0fb30fa7cca9718222038a2b578d40c4f21c5e350b74e753b85587"} Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.845132 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjk9p" event={"ID":"3b3fb69e-dd4f-4787-a207-4fe25106f9e7","Type":"ContainerStarted","Data":"7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06"} Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.845520 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.847571 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khr6h" event={"ID":"1d185732-cd6b-44c6-b4db-ee9ade00c683","Type":"ContainerStarted","Data":"b352e6584b478e7228a408cc5d6c8b18473e75a0de7be819c32ae9b98a707a4e"} Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.850042 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.852077 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.852570 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.852843 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.852946 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.853091 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.853423 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.853771 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.855075 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.855124 4778 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1" exitCode=0 Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.855390 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.855604 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.855780 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.855925 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.856135 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.856938 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae"} Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.857781 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.857927 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.858146 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.858718 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.859279 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.859795 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.860086 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.860603 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtjz5" event={"ID":"b9bef112-9bef-4ce2-abd8-054b4d671658","Type":"ContainerStarted","Data":"3151ddc8cb64182fd7ccd241e4580f2e0243328e43f1e59366f60b980b160490"} Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.861585 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.861764 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.861915 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.862067 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.862230 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.862372 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.862514 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: I0312 13:14:27.862654 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:27 crc kubenswrapper[4778]: E0312 13:14:27.912611 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="1.6s" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.918045 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.918830 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.919455 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.919738 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.920290 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.920982 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.921266 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.921517 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.921819 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.922209 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:28 crc kubenswrapper[4778]: I0312 13:14:28.922496 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.004659 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.004783 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.004811 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.004807 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.004902 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.004991 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.005351 4778 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.005371 4778 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.005381 4778 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:29 crc kubenswrapper[4778]: E0312 13:14:29.352195 4778 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.32:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" volumeName="registry-storage" Mar 12 13:14:29 crc kubenswrapper[4778]: E0312 13:14:29.513482 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="3.2s" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.884147 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.886104 4778 scope.go:117] "RemoveContainer" containerID="5019c5de667abecf425384b69c58060050b28003230e410f44934c9a7ad5484c" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.886223 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.901482 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.902112 4778 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.902451 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.903936 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.904488 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.904780 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.905024 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.905575 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.906040 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.908911 4778 scope.go:117] "RemoveContainer" containerID="bdfb81ab3f0178dc8064bd278e9e5cc42b3b2fda7282bb869d2f385b423e57d0" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.931551 4778 scope.go:117] "RemoveContainer" containerID="bc7259359df220c534d265305ee3ca44e7bcdce8da0d8b164132e02f7ed72e51" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.949347 4778 scope.go:117] "RemoveContainer" containerID="2d60adb329e51ce7d877de68c1386f904ef0f717c82a5bfb69ab18438a4e536a" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.965275 4778 scope.go:117] "RemoveContainer" containerID="7f640289dea724d5668fc009d628345ea104b2bbc9bc3471e42c3ec5f9acada1" Mar 12 13:14:29 crc kubenswrapper[4778]: I0312 13:14:29.979777 4778 scope.go:117] "RemoveContainer" containerID="e64aa9b1a15198d88b5f38b8ad0abdeef89430869b6f25c73e2f45806c539964" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.263961 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.892827 4778 generic.go:334] "Generic (PLEG): container finished" podID="a868c6a4-19ec-46be-a0af-be25b1049ff3" containerID="35dc89f42df73eafd54f7518d380b5b4f6934732de9c6dd0209b64b9345aa66c" exitCode=0 Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.892906 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a868c6a4-19ec-46be-a0af-be25b1049ff3","Type":"ContainerDied","Data":"35dc89f42df73eafd54f7518d380b5b4f6934732de9c6dd0209b64b9345aa66c"} Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.893450 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.893719 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.894049 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.894337 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.894587 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.894822 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.895029 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.895241 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:30 crc kubenswrapper[4778]: I0312 13:14:30.895457 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:31 crc kubenswrapper[4778]: E0312 13:14:31.611503 4778 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.32:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-sjk9p.189c1a4bbda7962d openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-sjk9p,UID:3b3fb69e-dd4f-4787-a207-4fe25106f9e7,APIVersion:v1,ResourceVersion:28313,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 16.256s (16.256s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 13:14:25.883878957 +0000 UTC m=+284.332574353,LastTimestamp:2026-03-12 13:14:25.883878957 +0000 UTC m=+284.332574353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.188275 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.188755 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.189068 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.189244 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.189393 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.189529 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.189671 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.189809 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.190006 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.190212 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.250342 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-var-lock\") pod \"a868c6a4-19ec-46be-a0af-be25b1049ff3\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.250444 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-var-lock" (OuterVolumeSpecName: "var-lock") pod "a868c6a4-19ec-46be-a0af-be25b1049ff3" (UID: "a868c6a4-19ec-46be-a0af-be25b1049ff3"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.250803 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a868c6a4-19ec-46be-a0af-be25b1049ff3-kube-api-access\") pod \"a868c6a4-19ec-46be-a0af-be25b1049ff3\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.250906 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-kubelet-dir\") pod \"a868c6a4-19ec-46be-a0af-be25b1049ff3\" (UID: \"a868c6a4-19ec-46be-a0af-be25b1049ff3\") " Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.250962 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a868c6a4-19ec-46be-a0af-be25b1049ff3" (UID: "a868c6a4-19ec-46be-a0af-be25b1049ff3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.251594 4778 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-var-lock\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.251655 4778 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a868c6a4-19ec-46be-a0af-be25b1049ff3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.255849 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a868c6a4-19ec-46be-a0af-be25b1049ff3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a868c6a4-19ec-46be-a0af-be25b1049ff3" (UID: "a868c6a4-19ec-46be-a0af-be25b1049ff3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.256959 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.257359 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.257578 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.257822 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.258223 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.258689 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.259009 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.259549 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.259754 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.352897 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a868c6a4-19ec-46be-a0af-be25b1049ff3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:32 crc kubenswrapper[4778]: E0312 13:14:32.714568 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="6.4s" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.907796 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a868c6a4-19ec-46be-a0af-be25b1049ff3","Type":"ContainerDied","Data":"67b06efe996403c5470e41a5f9a62a78fe522b551d7ec62d8302163676162a07"} Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.908134 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67b06efe996403c5470e41a5f9a62a78fe522b551d7ec62d8302163676162a07" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.907844 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.913419 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.913965 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.914412 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.914763 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.915234 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.915619 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.915954 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.916312 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:32 crc kubenswrapper[4778]: I0312 13:14:32.916774 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.693412 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.693497 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.754352 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.755133 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.755642 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.756045 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.756292 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.756568 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.756920 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.757204 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.757501 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.757830 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.863099 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.863164 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.913410 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.914068 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.914515 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.915121 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.915399 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.915749 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.916332 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.917043 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.917514 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.917875 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.959363 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.959841 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.960144 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.960539 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.960780 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.961015 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.961234 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.961297 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.961451 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.961694 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.961955 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.962316 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.962558 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.962808 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.963074 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.963322 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.963559 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.963777 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.964023 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:33 crc kubenswrapper[4778]: I0312 13:14:33.964335 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.020155 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.020239 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.056320 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.056952 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.057713 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.058404 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.058809 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.059260 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.059672 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.059996 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.060354 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.060791 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.309136 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.309230 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.355537 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.356073 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.356385 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.356737 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.356964 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.357216 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.357432 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.357642 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.357854 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.358075 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.972345 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.973104 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.973901 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.974464 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.974892 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.975406 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.975898 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.976283 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.976573 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.976995 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.989017 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.989748 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.990069 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.990705 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.991225 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.991760 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.992353 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.993095 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.993608 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:34 crc kubenswrapper[4778]: I0312 13:14:34.994258 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.024221 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.024305 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.088922 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.089889 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.090432 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.090995 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.091417 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.091746 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.092166 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.092646 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.092954 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.093472 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.980078 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.981015 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.981513 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.982119 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.982802 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.983148 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.983727 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.983979 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.984295 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:36 crc kubenswrapper[4778]: I0312 13:14:36.984858 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.243894 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:37Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:37Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:37Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T13:14:37Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:0d4c830b2653f2eeffebd09537afb06afb5ae827adbc03f224ab7269f399c05c\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:d6065909bc521a3f9a85174276fdbceafad02a276449a7dd1952a1f689b0d362\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1735807445},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:185237e125a9d710a58d4b588ea6b75eb361e4e99d979c1acd193de3b5d787f1\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:746054bb64fa0b27b1a696cd5db508bb9ee883a94969e4c1c4b5d35a93da8ef5\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1281521882},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:898c67bf7fc973e99114f3148976a6c21ae0dbe413051415588fa9b995f5b331\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:a641939d2096609a4cf6eec872a1476b7c671bfd81cffc2edeb6e9f13c9deeba\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1231028434},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0e6908b5c2800b56584a3fdf3bc164b76cb945966a49103123dabb61f8e367f2\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:ad31505e97766fe3b9d49abfe33098361de32a828c13e290be714f02a7ee76e0\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1221788890},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.244409 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.244632 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.244854 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.244999 4778 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.245012 4778 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.253362 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.254000 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.254705 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.255095 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.255579 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.256071 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.256533 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.256824 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.257116 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.257415 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.272931 4778 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.272958 4778 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:37 crc kubenswrapper[4778]: E0312 13:14:37.273571 4778 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.274496 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:37 crc kubenswrapper[4778]: W0312 13:14:37.293824 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-2024a3d05ac9cecb262ce575a431d92a7721c2423373b7acb5e57e5d294feecd WatchSource:0}: Error finding container 2024a3d05ac9cecb262ce575a431d92a7721c2423373b7acb5e57e5d294feecd: Status 404 returned error can't find the container with id 2024a3d05ac9cecb262ce575a431d92a7721c2423373b7acb5e57e5d294feecd Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.680347 4778 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.680784 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.958007 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2024a3d05ac9cecb262ce575a431d92a7721c2423373b7acb5e57e5d294feecd"} Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.961062 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.961951 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.962094 4778 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="62d772ee1ff9d986b4311494a08c8763bd91704fda6cd9c6f067c98205a4067d" exitCode=1 Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.962138 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"62d772ee1ff9d986b4311494a08c8763bd91704fda6cd9c6f067c98205a4067d"} Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.963497 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.963547 4778 scope.go:117] "RemoveContainer" containerID="62d772ee1ff9d986b4311494a08c8763bd91704fda6cd9c6f067c98205a4067d" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.963763 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.964016 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.964333 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.965018 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.965576 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.966404 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.967121 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.967748 4778 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:37 crc kubenswrapper[4778]: I0312 13:14:37.968159 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.420929 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.970457 4778 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c3ba0ec0d045f651180121b0ffc75036762171c599bf61f1dbb06d5e91189e14" exitCode=0 Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.970556 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c3ba0ec0d045f651180121b0ffc75036762171c599bf61f1dbb06d5e91189e14"} Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.970859 4778 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.970874 4778 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:38 crc kubenswrapper[4778]: E0312 13:14:38.971605 4778 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.971881 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.972392 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.972875 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.973156 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.973459 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.973830 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.974588 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.974686 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.975008 4778 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.975245 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.975292 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b062a3e9d7d1fd4f8252de53fc5e70585f0a129e0886c22103d39b30c3ecf110"} Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.975369 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.975621 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.975994 4778 status_manager.go:851] "Failed to get status for pod" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" pod="openshift-marketplace/certified-operators-qx9d8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-qx9d8\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.976298 4778 status_manager.go:851] "Failed to get status for pod" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" pod="openshift-marketplace/community-operators-khr6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-khr6h\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.976597 4778 status_manager.go:851] "Failed to get status for pod" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" pod="openshift-marketplace/certified-operators-l8n9b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-l8n9b\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.976931 4778 status_manager.go:851] "Failed to get status for pod" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" pod="openshift-marketplace/redhat-marketplace-rtjz5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rtjz5\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.977233 4778 status_manager.go:851] "Failed to get status for pod" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" pod="openshift-marketplace/redhat-marketplace-8xksl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8xksl\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.977528 4778 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.977764 4778 status_manager.go:851] "Failed to get status for pod" podUID="34ecd758-517c-455a-939a-7eb6d3546854" pod="openshift-marketplace/redhat-operators-76s88" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-76s88\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.978024 4778 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.978303 4778 status_manager.go:851] "Failed to get status for pod" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" pod="openshift-marketplace/community-operators-sjk9p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-sjk9p\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:38 crc kubenswrapper[4778]: I0312 13:14:38.978643 4778 status_manager.go:851] "Failed to get status for pod" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.32:6443: connect: connection refused" Mar 12 13:14:39 crc kubenswrapper[4778]: E0312 13:14:39.115198 4778 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.32:6443: connect: connection refused" interval="7s" Mar 12 13:14:39 crc kubenswrapper[4778]: I0312 13:14:39.985423 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"19c733da1c83d3d6ba0682c73ec41046fadd8031ec4303ee9495e0db3e977a6d"} Mar 12 13:14:39 crc kubenswrapper[4778]: I0312 13:14:39.985485 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0f85a6b014ee32c1af2cc5c218ef1c7fb4a60d752368d33b696855f34eac6c5b"} Mar 12 13:14:39 crc kubenswrapper[4778]: I0312 13:14:39.985497 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b63f23fa8170e58a310dd20337f65adb16b8d5342fa6ba2ed14f4ddc1fbd544c"} Mar 12 13:14:40 crc kubenswrapper[4778]: I0312 13:14:40.997041 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e4feca0a2e0663b03318b7fdaf2f14c43fb0c91c647627501b7d93e076309212"} Mar 12 13:14:40 crc kubenswrapper[4778]: I0312 13:14:40.997505 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"51cac27142c0b53681a48ca8823e0010131e54f580c2880e477a076b3a7108e9"} Mar 12 13:14:40 crc kubenswrapper[4778]: I0312 13:14:40.997523 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:40 crc kubenswrapper[4778]: I0312 13:14:40.997286 4778 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:40 crc kubenswrapper[4778]: I0312 13:14:40.997542 4778 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:41 crc kubenswrapper[4778]: I0312 13:14:41.850536 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" podUID="f36ec67c-df24-46ce-94b9-10619822c15a" containerName="oauth-openshift" containerID="cri-o://0fe97ea87ef2b2f3106d61689b8bc6549f4b603dd4e79e424ddbe8637587b2f3" gracePeriod=15 Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.007305 4778 generic.go:334] "Generic (PLEG): container finished" podID="f36ec67c-df24-46ce-94b9-10619822c15a" containerID="0fe97ea87ef2b2f3106d61689b8bc6549f4b603dd4e79e424ddbe8637587b2f3" exitCode=0 Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.007351 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" event={"ID":"f36ec67c-df24-46ce-94b9-10619822c15a","Type":"ContainerDied","Data":"0fe97ea87ef2b2f3106d61689b8bc6549f4b603dd4e79e424ddbe8637587b2f3"} Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.236780 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.274868 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.274909 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.280095 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299287 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-serving-cert\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299331 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-provider-selection\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299354 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-idp-0-file-data\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299376 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-router-certs\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299410 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-session\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299442 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-audit-policies\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299458 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-trusted-ca-bundle\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299479 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36ec67c-df24-46ce-94b9-10619822c15a-audit-dir\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299505 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-login\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299520 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-ocp-branding-template\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299540 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-error\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299556 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-cliconfig\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299589 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrlvl\" (UniqueName: \"kubernetes.io/projected/f36ec67c-df24-46ce-94b9-10619822c15a-kube-api-access-xrlvl\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.299630 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-service-ca\") pod \"f36ec67c-df24-46ce-94b9-10619822c15a\" (UID: \"f36ec67c-df24-46ce-94b9-10619822c15a\") " Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.301058 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.301204 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.301157 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.301411 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f36ec67c-df24-46ce-94b9-10619822c15a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.301519 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.305053 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f36ec67c-df24-46ce-94b9-10619822c15a-kube-api-access-xrlvl" (OuterVolumeSpecName: "kube-api-access-xrlvl") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "kube-api-access-xrlvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.305255 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.305350 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.305560 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.305922 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.306250 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.306442 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.306622 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.311561 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f36ec67c-df24-46ce-94b9-10619822c15a" (UID: "f36ec67c-df24-46ce-94b9-10619822c15a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400620 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400648 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400660 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400670 4778 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400679 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400688 4778 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36ec67c-df24-46ce-94b9-10619822c15a-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400698 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400707 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400716 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400724 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400733 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrlvl\" (UniqueName: \"kubernetes.io/projected/f36ec67c-df24-46ce-94b9-10619822c15a-kube-api-access-xrlvl\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400741 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400748 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:42 crc kubenswrapper[4778]: I0312 13:14:42.400757 4778 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f36ec67c-df24-46ce-94b9-10619822c15a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 12 13:14:43 crc kubenswrapper[4778]: I0312 13:14:43.015047 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" event={"ID":"f36ec67c-df24-46ce-94b9-10619822c15a","Type":"ContainerDied","Data":"5f7362fc7516f559081256deebf693613a994486c74f126dfda003689ad66bff"} Mar 12 13:14:43 crc kubenswrapper[4778]: I0312 13:14:43.016433 4778 scope.go:117] "RemoveContainer" containerID="0fe97ea87ef2b2f3106d61689b8bc6549f4b603dd4e79e424ddbe8637587b2f3" Mar 12 13:14:43 crc kubenswrapper[4778]: I0312 13:14:43.015098 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kw4v" Mar 12 13:14:46 crc kubenswrapper[4778]: I0312 13:14:46.006224 4778 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:46 crc kubenswrapper[4778]: I0312 13:14:46.038181 4778 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:46 crc kubenswrapper[4778]: I0312 13:14:46.038231 4778 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:46 crc kubenswrapper[4778]: I0312 13:14:46.044380 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:14:46 crc kubenswrapper[4778]: I0312 13:14:46.264961 4778 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="20ab8792-6c0f-49d4-993b-8fde07fc779e" Mar 12 13:14:46 crc kubenswrapper[4778]: E0312 13:14:46.911284 4778 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Mar 12 13:14:46 crc kubenswrapper[4778]: E0312 13:14:46.958618 4778 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Mar 12 13:14:47 crc kubenswrapper[4778]: I0312 13:14:47.052222 4778 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:47 crc kubenswrapper[4778]: I0312 13:14:47.052267 4778 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d15dec8c-5c3e-4103-a5b1-6ee7ff5990ef" Mar 12 13:14:47 crc kubenswrapper[4778]: I0312 13:14:47.057001 4778 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="20ab8792-6c0f-49d4-993b-8fde07fc779e" Mar 12 13:14:47 crc kubenswrapper[4778]: E0312 13:14:47.119607 4778 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Mar 12 13:14:47 crc kubenswrapper[4778]: I0312 13:14:47.203879 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:14:47 crc kubenswrapper[4778]: I0312 13:14:47.207396 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:14:48 crc kubenswrapper[4778]: I0312 13:14:48.059795 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:14:48 crc kubenswrapper[4778]: I0312 13:14:48.067005 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 13:14:56 crc kubenswrapper[4778]: I0312 13:14:56.281341 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 12 13:14:56 crc kubenswrapper[4778]: I0312 13:14:56.473658 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 12 13:14:56 crc kubenswrapper[4778]: I0312 13:14:56.703230 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 12 13:14:56 crc kubenswrapper[4778]: I0312 13:14:56.872372 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 12 13:14:56 crc kubenswrapper[4778]: I0312 13:14:56.932759 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.048687 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.143745 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.427767 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.463891 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.771172 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.849277 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.945423 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 12 13:14:57 crc kubenswrapper[4778]: I0312 13:14:57.947703 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.004015 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.220041 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.299932 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.674143 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.770169 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.891958 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.894893 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 12 13:14:58 crc kubenswrapper[4778]: I0312 13:14:58.897052 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.028112 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.061656 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.134006 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.157003 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.247125 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.282864 4778 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.323556 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.699876 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.796175 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.806427 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.829805 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.866175 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.937907 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.971970 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 12 13:14:59 crc kubenswrapper[4778]: I0312 13:14:59.993587 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.049848 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.050593 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.062028 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.063271 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.105936 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.164651 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.177436 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.334284 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.493623 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.593463 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.647337 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.662073 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.719570 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.875627 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.901346 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.932115 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 13:15:00 crc kubenswrapper[4778]: I0312 13:15:00.977587 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.038601 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.091131 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.102527 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.160030 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.173701 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.214123 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.268284 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.330272 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.432860 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.564158 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.583337 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.633617 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.669482 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.671506 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.688210 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.717745 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.749837 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.760734 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 12 13:15:01 crc kubenswrapper[4778]: I0312 13:15:01.940753 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.070431 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.074584 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.160301 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.229274 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.238227 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.299254 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.310097 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.358884 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.442282 4778 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.443068 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rtjz5" podStartSLOduration=38.407977804 podStartE2EDuration="2m7.443046296s" podCreationTimestamp="2026-03-12 13:12:55 +0000 UTC" firstStartedPulling="2026-03-12 13:12:57.064590534 +0000 UTC m=+195.513285930" lastFinishedPulling="2026-03-12 13:14:26.099658986 +0000 UTC m=+284.548354422" observedRunningTime="2026-03-12 13:14:46.061051499 +0000 UTC m=+304.509746915" watchObservedRunningTime="2026-03-12 13:15:02.443046296 +0000 UTC m=+320.891741692" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.444544 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qx9d8" podStartSLOduration=39.516571354 podStartE2EDuration="2m9.444533483s" podCreationTimestamp="2026-03-12 13:12:53 +0000 UTC" firstStartedPulling="2026-03-12 13:12:55.9649896 +0000 UTC m=+194.413684996" lastFinishedPulling="2026-03-12 13:14:25.892951739 +0000 UTC m=+284.341647125" observedRunningTime="2026-03-12 13:14:46.154097241 +0000 UTC m=+304.602792637" watchObservedRunningTime="2026-03-12 13:15:02.444533483 +0000 UTC m=+320.893228879" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.445794 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l8n9b" podStartSLOduration=42.824742371 podStartE2EDuration="2m9.445784314s" podCreationTimestamp="2026-03-12 13:12:53 +0000 UTC" firstStartedPulling="2026-03-12 13:12:55.827683035 +0000 UTC m=+194.276378431" lastFinishedPulling="2026-03-12 13:14:22.448724978 +0000 UTC m=+280.897420374" observedRunningTime="2026-03-12 13:14:46.028791911 +0000 UTC m=+304.477487327" watchObservedRunningTime="2026-03-12 13:15:02.445784314 +0000 UTC m=+320.894479710" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.447586 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.447574871 podStartE2EDuration="38.447574871s" podCreationTimestamp="2026-03-12 13:14:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:14:46.100152366 +0000 UTC m=+304.548847782" watchObservedRunningTime="2026-03-12 13:15:02.447574871 +0000 UTC m=+320.896270267" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.448045 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sjk9p" podStartSLOduration=39.463259805 podStartE2EDuration="2m9.448038256s" podCreationTimestamp="2026-03-12 13:12:53 +0000 UTC" firstStartedPulling="2026-03-12 13:12:55.899088745 +0000 UTC m=+194.347784141" lastFinishedPulling="2026-03-12 13:14:25.883867196 +0000 UTC m=+284.332562592" observedRunningTime="2026-03-12 13:14:46.113514356 +0000 UTC m=+304.562209792" watchObservedRunningTime="2026-03-12 13:15:02.448038256 +0000 UTC m=+320.896733652" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.449559 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-khr6h" podStartSLOduration=46.815371593 podStartE2EDuration="2m9.449548285s" podCreationTimestamp="2026-03-12 13:12:53 +0000 UTC" firstStartedPulling="2026-03-12 13:12:54.777716279 +0000 UTC m=+193.226411675" lastFinishedPulling="2026-03-12 13:14:17.411892951 +0000 UTC m=+275.860588367" observedRunningTime="2026-03-12 13:14:46.192175526 +0000 UTC m=+304.640870922" watchObservedRunningTime="2026-03-12 13:15:02.449548285 +0000 UTC m=+320.898243681" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.450624 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-76s88","openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-5kw4v"] Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.450697 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.459585 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.467853 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.474638 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.474618431 podStartE2EDuration="16.474618431s" podCreationTimestamp="2026-03-12 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:02.473065981 +0000 UTC m=+320.921761387" watchObservedRunningTime="2026-03-12 13:15:02.474618431 +0000 UTC m=+320.923313827" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.514171 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.528290 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.613572 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.684864 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.721527 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.741610 4778 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.753237 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.756249 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.804623 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.887318 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 12 13:15:02 crc kubenswrapper[4778]: I0312 13:15:02.897856 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.124927 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.229333 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.273931 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.381129 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.451412 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.475908 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.483352 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.591083 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.622649 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 12 13:15:03 crc kubenswrapper[4778]: I0312 13:15:03.914501 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.117501 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.260033 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ecd758-517c-455a-939a-7eb6d3546854" path="/var/lib/kubelet/pods/34ecd758-517c-455a-939a-7eb6d3546854/volumes" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.260690 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f36ec67c-df24-46ce-94b9-10619822c15a" path="/var/lib/kubelet/pods/f36ec67c-df24-46ce-94b9-10619822c15a/volumes" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.283265 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.316463 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.317149 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.327975 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.336266 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.500261 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.591923 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.641495 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.670605 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.672639 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.673939 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.695880 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.945122 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.985010 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.986520 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987058 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555355-68226"] Mar 12 13:15:04 crc kubenswrapper[4778]: E0312 13:15:04.987263 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="registry-server" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987280 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="registry-server" Mar 12 13:15:04 crc kubenswrapper[4778]: E0312 13:15:04.987291 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="extract-utilities" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987297 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="extract-utilities" Mar 12 13:15:04 crc kubenswrapper[4778]: E0312 13:15:04.987305 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="extract-content" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987312 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="extract-content" Mar 12 13:15:04 crc kubenswrapper[4778]: E0312 13:15:04.987327 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36ec67c-df24-46ce-94b9-10619822c15a" containerName="oauth-openshift" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987333 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36ec67c-df24-46ce-94b9-10619822c15a" containerName="oauth-openshift" Mar 12 13:15:04 crc kubenswrapper[4778]: E0312 13:15:04.987343 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" containerName="installer" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987348 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" containerName="installer" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987444 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36ec67c-df24-46ce-94b9-10619822c15a" containerName="oauth-openshift" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987455 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a868c6a4-19ec-46be-a0af-be25b1049ff3" containerName="installer" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987462 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ecd758-517c-455a-939a-7eb6d3546854" containerName="registry-server" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.987767 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.990045 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 13:15:04 crc kubenswrapper[4778]: I0312 13:15:04.990382 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.052860 4778 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.086969 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.116757 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25sw\" (UniqueName: \"kubernetes.io/projected/6197b3a9-f02f-4e5d-8196-b617fffa467d-kube-api-access-d25sw\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.116825 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6197b3a9-f02f-4e5d-8196-b617fffa467d-config-volume\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.116883 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6197b3a9-f02f-4e5d-8196-b617fffa467d-secret-volume\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.123936 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.182374 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.182862 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.196230 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.217665 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25sw\" (UniqueName: \"kubernetes.io/projected/6197b3a9-f02f-4e5d-8196-b617fffa467d-kube-api-access-d25sw\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.217753 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6197b3a9-f02f-4e5d-8196-b617fffa467d-config-volume\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.217819 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6197b3a9-f02f-4e5d-8196-b617fffa467d-secret-volume\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.219025 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6197b3a9-f02f-4e5d-8196-b617fffa467d-config-volume\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.225574 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.227327 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.229566 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6197b3a9-f02f-4e5d-8196-b617fffa467d-secret-volume\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.240780 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25sw\" (UniqueName: \"kubernetes.io/projected/6197b3a9-f02f-4e5d-8196-b617fffa467d-kube-api-access-d25sw\") pod \"collect-profiles-29555355-68226\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.251212 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.262571 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.286413 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.306515 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.314750 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.336764 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.454628 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.532713 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.534225 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.547714 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.583227 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.607835 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.705421 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.709387 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.781586 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.851522 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 12 13:15:05 crc kubenswrapper[4778]: I0312 13:15:05.974170 4778 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.019117 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.090089 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.102665 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.132343 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.207073 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.208023 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.279168 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.366515 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.366754 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.398644 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.474892 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.509130 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.513555 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.513709 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.519691 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.593888 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.841220 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.856310 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 13:15:06 crc kubenswrapper[4778]: I0312 13:15:06.940547 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.017275 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.035882 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.042266 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.302703 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.308866 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.321860 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.339412 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.340672 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.355332 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.419250 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.457774 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.555636 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.599140 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.641649 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx"] Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.642338 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.644240 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.644876 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.646628 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.646636 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.646905 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.647061 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.647688 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.647868 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.648168 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.648481 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.650462 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.650709 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.653498 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.656633 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.661304 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.695049 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.740056 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760064 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-error\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760123 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760151 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760173 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760235 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88f550ae-456c-496a-ae0c-e0e063022780-audit-dir\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760325 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-audit-policies\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760387 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-login\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760524 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760564 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760599 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-session\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760627 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760672 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760726 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.760765 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2pf8\" (UniqueName: \"kubernetes.io/projected/88f550ae-456c-496a-ae0c-e0e063022780-kube-api-access-f2pf8\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.800367 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.828387 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.830785 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.843474 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.861596 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862243 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-login\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862316 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862361 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862390 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-session\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862414 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862446 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862473 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862499 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2pf8\" (UniqueName: \"kubernetes.io/projected/88f550ae-456c-496a-ae0c-e0e063022780-kube-api-access-f2pf8\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862528 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-error\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862572 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862598 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862622 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862649 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88f550ae-456c-496a-ae0c-e0e063022780-audit-dir\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.862681 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-audit-policies\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.863450 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88f550ae-456c-496a-ae0c-e0e063022780-audit-dir\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.863653 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-audit-policies\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.863802 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.864298 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.865361 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.868864 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.869287 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.869477 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-session\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.869973 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.870243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.874228 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-login\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.875275 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.884458 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88f550ae-456c-496a-ae0c-e0e063022780-v4-0-config-user-template-error\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.885747 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2pf8\" (UniqueName: \"kubernetes.io/projected/88f550ae-456c-496a-ae0c-e0e063022780-kube-api-access-f2pf8\") pod \"oauth-openshift-7d48c8fbd6-kk9gx\" (UID: \"88f550ae-456c-496a-ae0c-e0e063022780\") " pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.919354 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.962481 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:07 crc kubenswrapper[4778]: I0312 13:15:07.995505 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.001630 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.058630 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.117885 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.117887 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.160870 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.177243 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.209722 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.341782 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.395466 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.438475 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.468852 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.500986 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.566111 4778 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.566578 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae" gracePeriod=5 Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.605007 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.631241 4778 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.682734 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.694905 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.734073 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.761100 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.908331 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.928941 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 12 13:15:08 crc kubenswrapper[4778]: I0312 13:15:08.963052 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.011794 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.194009 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.203312 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.374065 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.692534 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.951335 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:09 crc kubenswrapper[4778]: I0312 13:15:09.983441 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.344775 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.452657 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.469479 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.588857 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.588912 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.693055 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.732288 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.767768 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx"] Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.775824 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555355-68226"] Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.862505 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 13:15:10 crc kubenswrapper[4778]: I0312 13:15:10.896597 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.082119 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.208571 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555355-68226"] Mar 12 13:15:11 crc kubenswrapper[4778]: W0312 13:15:11.212133 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6197b3a9_f02f_4e5d_8196_b617fffa467d.slice/crio-9fef4c59e32339bdfc08c1427d6779038e13657c03fafb52c99c44018f2fa182 WatchSource:0}: Error finding container 9fef4c59e32339bdfc08c1427d6779038e13657c03fafb52c99c44018f2fa182: Status 404 returned error can't find the container with id 9fef4c59e32339bdfc08c1427d6779038e13657c03fafb52c99c44018f2fa182 Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.255553 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.263500 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx"] Mar 12 13:15:11 crc kubenswrapper[4778]: W0312 13:15:11.268656 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88f550ae_456c_496a_ae0c_e0e063022780.slice/crio-d11163752c2ad0bf906f93a493f49241d8588cc59428d3f435eb62ef394d3b28 WatchSource:0}: Error finding container d11163752c2ad0bf906f93a493f49241d8588cc59428d3f435eb62ef394d3b28: Status 404 returned error can't find the container with id d11163752c2ad0bf906f93a493f49241d8588cc59428d3f435eb62ef394d3b28 Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.291536 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.417160 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.444606 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.611786 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.706603 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.877221 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.891328 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 12 13:15:11 crc kubenswrapper[4778]: I0312 13:15:11.933409 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.063720 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.204887 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" event={"ID":"88f550ae-456c-496a-ae0c-e0e063022780","Type":"ContainerStarted","Data":"bd045c8915afd2aafd5594b3924558bebda0349833512a1c5201d36a65cbcfed"} Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.205127 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" event={"ID":"88f550ae-456c-496a-ae0c-e0e063022780","Type":"ContainerStarted","Data":"d11163752c2ad0bf906f93a493f49241d8588cc59428d3f435eb62ef394d3b28"} Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.205156 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.208274 4778 generic.go:334] "Generic (PLEG): container finished" podID="6197b3a9-f02f-4e5d-8196-b617fffa467d" containerID="3954f4afdb430b04a44fc16681134a45669f465399452c67b26950fbb78cb40a" exitCode=0 Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.208343 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" event={"ID":"6197b3a9-f02f-4e5d-8196-b617fffa467d","Type":"ContainerDied","Data":"3954f4afdb430b04a44fc16681134a45669f465399452c67b26950fbb78cb40a"} Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.208387 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" event={"ID":"6197b3a9-f02f-4e5d-8196-b617fffa467d","Type":"ContainerStarted","Data":"9fef4c59e32339bdfc08c1427d6779038e13657c03fafb52c99c44018f2fa182"} Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.211367 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.216548 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.258787 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d48c8fbd6-kk9gx" podStartSLOduration=56.258768095 podStartE2EDuration="56.258768095s" podCreationTimestamp="2026-03-12 13:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:12.233578217 +0000 UTC m=+330.682273753" watchObservedRunningTime="2026-03-12 13:15:12.258768095 +0000 UTC m=+330.707463491" Mar 12 13:15:12 crc kubenswrapper[4778]: I0312 13:15:12.485658 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.437090 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.514699 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.537033 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25sw\" (UniqueName: \"kubernetes.io/projected/6197b3a9-f02f-4e5d-8196-b617fffa467d-kube-api-access-d25sw\") pod \"6197b3a9-f02f-4e5d-8196-b617fffa467d\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.537366 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6197b3a9-f02f-4e5d-8196-b617fffa467d-secret-volume\") pod \"6197b3a9-f02f-4e5d-8196-b617fffa467d\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.537474 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6197b3a9-f02f-4e5d-8196-b617fffa467d-config-volume\") pod \"6197b3a9-f02f-4e5d-8196-b617fffa467d\" (UID: \"6197b3a9-f02f-4e5d-8196-b617fffa467d\") " Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.537938 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6197b3a9-f02f-4e5d-8196-b617fffa467d-config-volume" (OuterVolumeSpecName: "config-volume") pod "6197b3a9-f02f-4e5d-8196-b617fffa467d" (UID: "6197b3a9-f02f-4e5d-8196-b617fffa467d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.542173 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6197b3a9-f02f-4e5d-8196-b617fffa467d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6197b3a9-f02f-4e5d-8196-b617fffa467d" (UID: "6197b3a9-f02f-4e5d-8196-b617fffa467d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.550095 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6197b3a9-f02f-4e5d-8196-b617fffa467d-kube-api-access-d25sw" (OuterVolumeSpecName: "kube-api-access-d25sw") pod "6197b3a9-f02f-4e5d-8196-b617fffa467d" (UID: "6197b3a9-f02f-4e5d-8196-b617fffa467d"). InnerVolumeSpecName "kube-api-access-d25sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.638317 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6197b3a9-f02f-4e5d-8196-b617fffa467d-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.638365 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25sw\" (UniqueName: \"kubernetes.io/projected/6197b3a9-f02f-4e5d-8196-b617fffa467d-kube-api-access-d25sw\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.638381 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6197b3a9-f02f-4e5d-8196-b617fffa467d-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.818326 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d44c8b88d-jx574"] Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.818866 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" podUID="0fdd5690-0e80-4317-9e3a-8478f09ea1a8" containerName="controller-manager" containerID="cri-o://de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398" gracePeriod=30 Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.926426 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5"] Mar 12 13:15:13 crc kubenswrapper[4778]: I0312 13:15:13.928550 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" podUID="badeb3df-9c56-4aa2-af6f-aba14c213fcc" containerName="route-controller-manager" containerID="cri-o://6edfc1174eae36c4699c23b09d94a6801a70e404d52b1e50d4350988d1f6d371" gracePeriod=30 Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.008590 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.115400 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.115641 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.212172 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.219692 4778 generic.go:334] "Generic (PLEG): container finished" podID="badeb3df-9c56-4aa2-af6f-aba14c213fcc" containerID="6edfc1174eae36c4699c23b09d94a6801a70e404d52b1e50d4350988d1f6d371" exitCode=0 Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.219796 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" event={"ID":"badeb3df-9c56-4aa2-af6f-aba14c213fcc","Type":"ContainerDied","Data":"6edfc1174eae36c4699c23b09d94a6801a70e404d52b1e50d4350988d1f6d371"} Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.221002 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" event={"ID":"6197b3a9-f02f-4e5d-8196-b617fffa467d","Type":"ContainerDied","Data":"9fef4c59e32339bdfc08c1427d6779038e13657c03fafb52c99c44018f2fa182"} Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.221024 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555355-68226" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.221031 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fef4c59e32339bdfc08c1427d6779038e13657c03fafb52c99c44018f2fa182" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.222487 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.222567 4778 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae" exitCode=137 Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.222642 4778 scope.go:117] "RemoveContainer" containerID="3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.222746 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.225821 4778 generic.go:334] "Generic (PLEG): container finished" podID="0fdd5690-0e80-4317-9e3a-8478f09ea1a8" containerID="de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398" exitCode=0 Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.225952 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" event={"ID":"0fdd5690-0e80-4317-9e3a-8478f09ea1a8","Type":"ContainerDied","Data":"de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398"} Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.225993 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" event={"ID":"0fdd5690-0e80-4317-9e3a-8478f09ea1a8","Type":"ContainerDied","Data":"dd0b16a4e92ddcc0e1151ae83fdba0245e8931b7997e602261eaa93e0a982440"} Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.226039 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d44c8b88d-jx574" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.243414 4778 scope.go:117] "RemoveContainer" containerID="3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae" Mar 12 13:15:14 crc kubenswrapper[4778]: E0312 13:15:14.243914 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae\": container with ID starting with 3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae not found: ID does not exist" containerID="3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.243951 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae"} err="failed to get container status \"3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae\": rpc error: code = NotFound desc = could not find container \"3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae\": container with ID starting with 3d740724bfd8227fe2e07ff4fe5fbe18790f3387faf339232729dc31b3dd39ae not found: ID does not exist" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.243976 4778 scope.go:117] "RemoveContainer" containerID="de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246166 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246235 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246261 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-client-ca\") pod \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246287 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-serving-cert\") pod \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246307 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246339 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-proxy-ca-bundles\") pod \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246370 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246387 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246408 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246420 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246429 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246444 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-config\") pod \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246469 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwdxl\" (UniqueName: \"kubernetes.io/projected/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-kube-api-access-zwdxl\") pod \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\" (UID: \"0fdd5690-0e80-4317-9e3a-8478f09ea1a8\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246841 4778 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246876 4778 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.246890 4778 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.247126 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-client-ca" (OuterVolumeSpecName: "client-ca") pod "0fdd5690-0e80-4317-9e3a-8478f09ea1a8" (UID: "0fdd5690-0e80-4317-9e3a-8478f09ea1a8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.248950 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.250859 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-kube-api-access-zwdxl" (OuterVolumeSpecName: "kube-api-access-zwdxl") pod "0fdd5690-0e80-4317-9e3a-8478f09ea1a8" (UID: "0fdd5690-0e80-4317-9e3a-8478f09ea1a8"). InnerVolumeSpecName "kube-api-access-zwdxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.251105 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-config" (OuterVolumeSpecName: "config") pod "0fdd5690-0e80-4317-9e3a-8478f09ea1a8" (UID: "0fdd5690-0e80-4317-9e3a-8478f09ea1a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.251725 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0fdd5690-0e80-4317-9e3a-8478f09ea1a8" (UID: "0fdd5690-0e80-4317-9e3a-8478f09ea1a8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.252149 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0fdd5690-0e80-4317-9e3a-8478f09ea1a8" (UID: "0fdd5690-0e80-4317-9e3a-8478f09ea1a8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.255740 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.256840 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.263510 4778 scope.go:117] "RemoveContainer" containerID="de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398" Mar 12 13:15:14 crc kubenswrapper[4778]: E0312 13:15:14.263928 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398\": container with ID starting with de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398 not found: ID does not exist" containerID="de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.263972 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398"} err="failed to get container status \"de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398\": rpc error: code = NotFound desc = could not find container \"de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398\": container with ID starting with de9da8336c3c506af0ce9ebe2cdd9483aff7c9248c270e308d85726473f6d398 not found: ID does not exist" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.264714 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.265029 4778 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.288788 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.288829 4778 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="11f6fe78-f714-49b6-ba0c-e07eefedd97e" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.291324 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.291372 4778 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="11f6fe78-f714-49b6-ba0c-e07eefedd97e" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347064 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-config\") pod \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347351 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/badeb3df-9c56-4aa2-af6f-aba14c213fcc-serving-cert\") pod \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347399 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-client-ca\") pod \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347426 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcsqr\" (UniqueName: \"kubernetes.io/projected/badeb3df-9c56-4aa2-af6f-aba14c213fcc-kube-api-access-tcsqr\") pod \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\" (UID: \"badeb3df-9c56-4aa2-af6f-aba14c213fcc\") " Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347680 4778 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347695 4778 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347707 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347718 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwdxl\" (UniqueName: \"kubernetes.io/projected/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-kube-api-access-zwdxl\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347730 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347742 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.347753 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fdd5690-0e80-4317-9e3a-8478f09ea1a8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.348639 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-client-ca" (OuterVolumeSpecName: "client-ca") pod "badeb3df-9c56-4aa2-af6f-aba14c213fcc" (UID: "badeb3df-9c56-4aa2-af6f-aba14c213fcc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.348664 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-config" (OuterVolumeSpecName: "config") pod "badeb3df-9c56-4aa2-af6f-aba14c213fcc" (UID: "badeb3df-9c56-4aa2-af6f-aba14c213fcc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.351436 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badeb3df-9c56-4aa2-af6f-aba14c213fcc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "badeb3df-9c56-4aa2-af6f-aba14c213fcc" (UID: "badeb3df-9c56-4aa2-af6f-aba14c213fcc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.351548 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/badeb3df-9c56-4aa2-af6f-aba14c213fcc-kube-api-access-tcsqr" (OuterVolumeSpecName: "kube-api-access-tcsqr") pod "badeb3df-9c56-4aa2-af6f-aba14c213fcc" (UID: "badeb3df-9c56-4aa2-af6f-aba14c213fcc"). InnerVolumeSpecName "kube-api-access-tcsqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.448723 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.448780 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/badeb3df-9c56-4aa2-af6f-aba14c213fcc-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.448796 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/badeb3df-9c56-4aa2-af6f-aba14c213fcc-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.448815 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcsqr\" (UniqueName: \"kubernetes.io/projected/badeb3df-9c56-4aa2-af6f-aba14c213fcc-kube-api-access-tcsqr\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.543375 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d44c8b88d-jx574"] Mar 12 13:15:14 crc kubenswrapper[4778]: I0312 13:15:14.547358 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-d44c8b88d-jx574"] Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.232648 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.232608 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5" event={"ID":"badeb3df-9c56-4aa2-af6f-aba14c213fcc","Type":"ContainerDied","Data":"ad682c14c40bcfdf45bab0f4aae014cbaeacf6b49ffe857ff368861fb7bbc412"} Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.232913 4778 scope.go:117] "RemoveContainer" containerID="6edfc1174eae36c4699c23b09d94a6801a70e404d52b1e50d4350988d1f6d371" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.263451 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5"] Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.272531 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f4dd5cc6-ppsx5"] Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.652529 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w"] Mar 12 13:15:15 crc kubenswrapper[4778]: E0312 13:15:15.652976 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="badeb3df-9c56-4aa2-af6f-aba14c213fcc" containerName="route-controller-manager" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.652987 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="badeb3df-9c56-4aa2-af6f-aba14c213fcc" containerName="route-controller-manager" Mar 12 13:15:15 crc kubenswrapper[4778]: E0312 13:15:15.653007 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fdd5690-0e80-4317-9e3a-8478f09ea1a8" containerName="controller-manager" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653013 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fdd5690-0e80-4317-9e3a-8478f09ea1a8" containerName="controller-manager" Mar 12 13:15:15 crc kubenswrapper[4778]: E0312 13:15:15.653022 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653029 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 13:15:15 crc kubenswrapper[4778]: E0312 13:15:15.653037 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6197b3a9-f02f-4e5d-8196-b617fffa467d" containerName="collect-profiles" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653043 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="6197b3a9-f02f-4e5d-8196-b617fffa467d" containerName="collect-profiles" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653151 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="6197b3a9-f02f-4e5d-8196-b617fffa467d" containerName="collect-profiles" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653166 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653173 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fdd5690-0e80-4317-9e3a-8478f09ea1a8" containerName="controller-manager" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653196 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="badeb3df-9c56-4aa2-af6f-aba14c213fcc" containerName="route-controller-manager" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653469 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt"] Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.653922 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.654282 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.659972 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.661787 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.665477 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.665825 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.666046 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.668818 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.668948 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.669635 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.669710 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.669908 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.670058 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.675053 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.684223 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.686065 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w"] Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.692726 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt"] Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765050 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-config\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765106 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctmx8\" (UniqueName: \"kubernetes.io/projected/e54ca63f-0568-4b0f-aa00-a726ead780cc-kube-api-access-ctmx8\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765131 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d6bd11-6c53-46a2-a38f-65672b1cc83f-serving-cert\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765147 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-proxy-ca-bundles\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765166 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ca63f-0568-4b0f-aa00-a726ead780cc-serving-cert\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765266 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-client-ca\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765287 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-config\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765462 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-client-ca\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.765555 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgrrw\" (UniqueName: \"kubernetes.io/projected/97d6bd11-6c53-46a2-a38f-65672b1cc83f-kube-api-access-cgrrw\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866573 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-client-ca\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866632 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-config\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866660 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-client-ca\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866703 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgrrw\" (UniqueName: \"kubernetes.io/projected/97d6bd11-6c53-46a2-a38f-65672b1cc83f-kube-api-access-cgrrw\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866760 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-config\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866819 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctmx8\" (UniqueName: \"kubernetes.io/projected/e54ca63f-0568-4b0f-aa00-a726ead780cc-kube-api-access-ctmx8\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866854 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d6bd11-6c53-46a2-a38f-65672b1cc83f-serving-cert\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866907 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-proxy-ca-bundles\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.866937 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ca63f-0568-4b0f-aa00-a726ead780cc-serving-cert\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.867811 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-client-ca\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.868275 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-config\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.868645 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-proxy-ca-bundles\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.868954 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-client-ca\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.869408 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-config\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.874416 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d6bd11-6c53-46a2-a38f-65672b1cc83f-serving-cert\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.874859 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ca63f-0568-4b0f-aa00-a726ead780cc-serving-cert\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.884309 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctmx8\" (UniqueName: \"kubernetes.io/projected/e54ca63f-0568-4b0f-aa00-a726ead780cc-kube-api-access-ctmx8\") pod \"controller-manager-7cd7bcbb47-r52bt\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.885565 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgrrw\" (UniqueName: \"kubernetes.io/projected/97d6bd11-6c53-46a2-a38f-65672b1cc83f-kube-api-access-cgrrw\") pod \"route-controller-manager-5d7995cfc7-rwc4w\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.975620 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:15 crc kubenswrapper[4778]: I0312 13:15:15.986840 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:16 crc kubenswrapper[4778]: I0312 13:15:16.211760 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt"] Mar 12 13:15:16 crc kubenswrapper[4778]: W0312 13:15:16.215855 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode54ca63f_0568_4b0f_aa00_a726ead780cc.slice/crio-e1f5fdc0207e5f640c33e64d7efe86ecc6a21b6930d23ea0075c4c50accc4ee6 WatchSource:0}: Error finding container e1f5fdc0207e5f640c33e64d7efe86ecc6a21b6930d23ea0075c4c50accc4ee6: Status 404 returned error can't find the container with id e1f5fdc0207e5f640c33e64d7efe86ecc6a21b6930d23ea0075c4c50accc4ee6 Mar 12 13:15:16 crc kubenswrapper[4778]: I0312 13:15:16.249777 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" event={"ID":"e54ca63f-0568-4b0f-aa00-a726ead780cc","Type":"ContainerStarted","Data":"e1f5fdc0207e5f640c33e64d7efe86ecc6a21b6930d23ea0075c4c50accc4ee6"} Mar 12 13:15:16 crc kubenswrapper[4778]: W0312 13:15:16.256692 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97d6bd11_6c53_46a2_a38f_65672b1cc83f.slice/crio-5047efe3d4d74f46dcbcd8ccb18595521f1d36baca9c39a378fbb3fd6752f393 WatchSource:0}: Error finding container 5047efe3d4d74f46dcbcd8ccb18595521f1d36baca9c39a378fbb3fd6752f393: Status 404 returned error can't find the container with id 5047efe3d4d74f46dcbcd8ccb18595521f1d36baca9c39a378fbb3fd6752f393 Mar 12 13:15:16 crc kubenswrapper[4778]: I0312 13:15:16.259079 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fdd5690-0e80-4317-9e3a-8478f09ea1a8" path="/var/lib/kubelet/pods/0fdd5690-0e80-4317-9e3a-8478f09ea1a8/volumes" Mar 12 13:15:16 crc kubenswrapper[4778]: I0312 13:15:16.260034 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="badeb3df-9c56-4aa2-af6f-aba14c213fcc" path="/var/lib/kubelet/pods/badeb3df-9c56-4aa2-af6f-aba14c213fcc/volumes" Mar 12 13:15:16 crc kubenswrapper[4778]: I0312 13:15:16.260558 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w"] Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.256913 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" event={"ID":"97d6bd11-6c53-46a2-a38f-65672b1cc83f","Type":"ContainerStarted","Data":"3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e"} Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.256952 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" event={"ID":"97d6bd11-6c53-46a2-a38f-65672b1cc83f","Type":"ContainerStarted","Data":"5047efe3d4d74f46dcbcd8ccb18595521f1d36baca9c39a378fbb3fd6752f393"} Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.258020 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.261654 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" event={"ID":"e54ca63f-0568-4b0f-aa00-a726ead780cc","Type":"ContainerStarted","Data":"32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386"} Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.262597 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.264017 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.268680 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.286845 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" podStartSLOduration=4.286826355 podStartE2EDuration="4.286826355s" podCreationTimestamp="2026-03-12 13:15:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:17.284648918 +0000 UTC m=+335.733344334" watchObservedRunningTime="2026-03-12 13:15:17.286826355 +0000 UTC m=+335.735521771" Mar 12 13:15:17 crc kubenswrapper[4778]: I0312 13:15:17.300467 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" podStartSLOduration=4.30044934 podStartE2EDuration="4.30044934s" podCreationTimestamp="2026-03-12 13:15:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:17.300298775 +0000 UTC m=+335.748994171" watchObservedRunningTime="2026-03-12 13:15:17.30044934 +0000 UTC m=+335.749144736" Mar 12 13:15:28 crc kubenswrapper[4778]: I0312 13:15:28.195149 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 12 13:15:49 crc kubenswrapper[4778]: I0312 13:15:49.528698 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt"] Mar 12 13:15:49 crc kubenswrapper[4778]: I0312 13:15:49.529440 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" podUID="e54ca63f-0568-4b0f-aa00-a726ead780cc" containerName="controller-manager" containerID="cri-o://32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386" gracePeriod=30 Mar 12 13:15:49 crc kubenswrapper[4778]: I0312 13:15:49.654008 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w"] Mar 12 13:15:49 crc kubenswrapper[4778]: I0312 13:15:49.654360 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" podUID="97d6bd11-6c53-46a2-a38f-65672b1cc83f" containerName="route-controller-manager" containerID="cri-o://3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e" gracePeriod=30 Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.019121 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.022816 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066146 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctmx8\" (UniqueName: \"kubernetes.io/projected/e54ca63f-0568-4b0f-aa00-a726ead780cc-kube-api-access-ctmx8\") pod \"e54ca63f-0568-4b0f-aa00-a726ead780cc\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066251 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d6bd11-6c53-46a2-a38f-65672b1cc83f-serving-cert\") pod \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066321 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgrrw\" (UniqueName: \"kubernetes.io/projected/97d6bd11-6c53-46a2-a38f-65672b1cc83f-kube-api-access-cgrrw\") pod \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066354 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-config\") pod \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066400 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ca63f-0568-4b0f-aa00-a726ead780cc-serving-cert\") pod \"e54ca63f-0568-4b0f-aa00-a726ead780cc\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066428 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-client-ca\") pod \"e54ca63f-0568-4b0f-aa00-a726ead780cc\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066471 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-client-ca\") pod \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\" (UID: \"97d6bd11-6c53-46a2-a38f-65672b1cc83f\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066533 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-proxy-ca-bundles\") pod \"e54ca63f-0568-4b0f-aa00-a726ead780cc\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.066570 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-config\") pod \"e54ca63f-0568-4b0f-aa00-a726ead780cc\" (UID: \"e54ca63f-0568-4b0f-aa00-a726ead780cc\") " Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.067706 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-config" (OuterVolumeSpecName: "config") pod "e54ca63f-0568-4b0f-aa00-a726ead780cc" (UID: "e54ca63f-0568-4b0f-aa00-a726ead780cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.067967 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-client-ca" (OuterVolumeSpecName: "client-ca") pod "e54ca63f-0568-4b0f-aa00-a726ead780cc" (UID: "e54ca63f-0568-4b0f-aa00-a726ead780cc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.068334 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-client-ca" (OuterVolumeSpecName: "client-ca") pod "97d6bd11-6c53-46a2-a38f-65672b1cc83f" (UID: "97d6bd11-6c53-46a2-a38f-65672b1cc83f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.068729 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e54ca63f-0568-4b0f-aa00-a726ead780cc" (UID: "e54ca63f-0568-4b0f-aa00-a726ead780cc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.068898 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-config" (OuterVolumeSpecName: "config") pod "97d6bd11-6c53-46a2-a38f-65672b1cc83f" (UID: "97d6bd11-6c53-46a2-a38f-65672b1cc83f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.072030 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e54ca63f-0568-4b0f-aa00-a726ead780cc-kube-api-access-ctmx8" (OuterVolumeSpecName: "kube-api-access-ctmx8") pod "e54ca63f-0568-4b0f-aa00-a726ead780cc" (UID: "e54ca63f-0568-4b0f-aa00-a726ead780cc"). InnerVolumeSpecName "kube-api-access-ctmx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.073471 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e54ca63f-0568-4b0f-aa00-a726ead780cc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e54ca63f-0568-4b0f-aa00-a726ead780cc" (UID: "e54ca63f-0568-4b0f-aa00-a726ead780cc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.074237 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97d6bd11-6c53-46a2-a38f-65672b1cc83f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "97d6bd11-6c53-46a2-a38f-65672b1cc83f" (UID: "97d6bd11-6c53-46a2-a38f-65672b1cc83f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.074871 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97d6bd11-6c53-46a2-a38f-65672b1cc83f-kube-api-access-cgrrw" (OuterVolumeSpecName: "kube-api-access-cgrrw") pod "97d6bd11-6c53-46a2-a38f-65672b1cc83f" (UID: "97d6bd11-6c53-46a2-a38f-65672b1cc83f"). InnerVolumeSpecName "kube-api-access-cgrrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167849 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167890 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167904 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctmx8\" (UniqueName: \"kubernetes.io/projected/e54ca63f-0568-4b0f-aa00-a726ead780cc-kube-api-access-ctmx8\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167914 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d6bd11-6c53-46a2-a38f-65672b1cc83f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167925 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgrrw\" (UniqueName: \"kubernetes.io/projected/97d6bd11-6c53-46a2-a38f-65672b1cc83f-kube-api-access-cgrrw\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167933 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167942 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ca63f-0568-4b0f-aa00-a726ead780cc-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167951 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e54ca63f-0568-4b0f-aa00-a726ead780cc-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.167959 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d6bd11-6c53-46a2-a38f-65672b1cc83f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.446462 4778 generic.go:334] "Generic (PLEG): container finished" podID="e54ca63f-0568-4b0f-aa00-a726ead780cc" containerID="32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386" exitCode=0 Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.446564 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.447165 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" event={"ID":"e54ca63f-0568-4b0f-aa00-a726ead780cc","Type":"ContainerDied","Data":"32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386"} Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.447217 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt" event={"ID":"e54ca63f-0568-4b0f-aa00-a726ead780cc","Type":"ContainerDied","Data":"e1f5fdc0207e5f640c33e64d7efe86ecc6a21b6930d23ea0075c4c50accc4ee6"} Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.447251 4778 scope.go:117] "RemoveContainer" containerID="32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.449912 4778 generic.go:334] "Generic (PLEG): container finished" podID="97d6bd11-6c53-46a2-a38f-65672b1cc83f" containerID="3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e" exitCode=0 Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.449945 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" event={"ID":"97d6bd11-6c53-46a2-a38f-65672b1cc83f","Type":"ContainerDied","Data":"3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e"} Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.449968 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" event={"ID":"97d6bd11-6c53-46a2-a38f-65672b1cc83f","Type":"ContainerDied","Data":"5047efe3d4d74f46dcbcd8ccb18595521f1d36baca9c39a378fbb3fd6752f393"} Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.450015 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.489910 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt"] Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.494935 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7cd7bcbb47-r52bt"] Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.504621 4778 scope.go:117] "RemoveContainer" containerID="32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386" Mar 12 13:15:50 crc kubenswrapper[4778]: E0312 13:15:50.505080 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386\": container with ID starting with 32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386 not found: ID does not exist" containerID="32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.505129 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386"} err="failed to get container status \"32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386\": rpc error: code = NotFound desc = could not find container \"32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386\": container with ID starting with 32e23f7128debfc4cc11679362c8383eacd2674f3f8f68c5993038e873d31386 not found: ID does not exist" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.505162 4778 scope.go:117] "RemoveContainer" containerID="3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.514297 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w"] Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.515059 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d7995cfc7-rwc4w"] Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.526598 4778 scope.go:117] "RemoveContainer" containerID="3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e" Mar 12 13:15:50 crc kubenswrapper[4778]: E0312 13:15:50.527009 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e\": container with ID starting with 3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e not found: ID does not exist" containerID="3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.527053 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e"} err="failed to get container status \"3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e\": rpc error: code = NotFound desc = could not find container \"3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e\": container with ID starting with 3f3381b2bee5e33ad3fd2e7b670b2802903ebb1e5d72d9179b8c4f23bd13d27e not found: ID does not exist" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.676270 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6"] Mar 12 13:15:50 crc kubenswrapper[4778]: E0312 13:15:50.676640 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d6bd11-6c53-46a2-a38f-65672b1cc83f" containerName="route-controller-manager" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.676671 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d6bd11-6c53-46a2-a38f-65672b1cc83f" containerName="route-controller-manager" Mar 12 13:15:50 crc kubenswrapper[4778]: E0312 13:15:50.676726 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e54ca63f-0568-4b0f-aa00-a726ead780cc" containerName="controller-manager" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.676743 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e54ca63f-0568-4b0f-aa00-a726ead780cc" containerName="controller-manager" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.676962 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="97d6bd11-6c53-46a2-a38f-65672b1cc83f" containerName="route-controller-manager" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.676989 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e54ca63f-0568-4b0f-aa00-a726ead780cc" containerName="controller-manager" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.677922 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.680493 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.680592 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.680980 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.681091 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.682065 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.684989 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.685910 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6"] Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.689938 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.776362 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8083240-e16c-40da-9f87-9db45cebfafd-serving-cert\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.776459 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-proxy-ca-bundles\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.776500 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jktc4\" (UniqueName: \"kubernetes.io/projected/c8083240-e16c-40da-9f87-9db45cebfafd-kube-api-access-jktc4\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.776530 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-config\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.776606 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-client-ca\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.877674 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-proxy-ca-bundles\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.877728 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jktc4\" (UniqueName: \"kubernetes.io/projected/c8083240-e16c-40da-9f87-9db45cebfafd-kube-api-access-jktc4\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.877752 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-config\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.877787 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-client-ca\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.877830 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8083240-e16c-40da-9f87-9db45cebfafd-serving-cert\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.878942 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-client-ca\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.879134 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-config\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.879856 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-proxy-ca-bundles\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.883946 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8083240-e16c-40da-9f87-9db45cebfafd-serving-cert\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.895838 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jktc4\" (UniqueName: \"kubernetes.io/projected/c8083240-e16c-40da-9f87-9db45cebfafd-kube-api-access-jktc4\") pod \"controller-manager-f8c4b6bf8-n4jk6\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:50 crc kubenswrapper[4778]: I0312 13:15:50.991693 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.406977 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6"] Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.461472 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" event={"ID":"c8083240-e16c-40da-9f87-9db45cebfafd","Type":"ContainerStarted","Data":"610f9e9935be3de3cd47d5e81cba88f5a4f3f12d0d389042b1b5255df2fa7476"} Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.689807 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r"] Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.690721 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.692738 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.692780 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.692784 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.695080 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r"] Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.695558 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.695759 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.695778 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.792339 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6eebe99-8b82-4720-82c5-c940100859ad-serving-cert\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.792406 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-client-ca\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.792475 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-config\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.792506 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc82s\" (UniqueName: \"kubernetes.io/projected/f6eebe99-8b82-4720-82c5-c940100859ad-kube-api-access-qc82s\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.893775 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6eebe99-8b82-4720-82c5-c940100859ad-serving-cert\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.893825 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-client-ca\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.893875 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-config\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.893903 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc82s\" (UniqueName: \"kubernetes.io/projected/f6eebe99-8b82-4720-82c5-c940100859ad-kube-api-access-qc82s\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.894785 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-client-ca\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.895310 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-config\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.898773 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6eebe99-8b82-4720-82c5-c940100859ad-serving-cert\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:51 crc kubenswrapper[4778]: I0312 13:15:51.916679 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc82s\" (UniqueName: \"kubernetes.io/projected/f6eebe99-8b82-4720-82c5-c940100859ad-kube-api-access-qc82s\") pod \"route-controller-manager-6d496c4846-ggb6r\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.052879 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.268487 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97d6bd11-6c53-46a2-a38f-65672b1cc83f" path="/var/lib/kubelet/pods/97d6bd11-6c53-46a2-a38f-65672b1cc83f/volumes" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.269725 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e54ca63f-0568-4b0f-aa00-a726ead780cc" path="/var/lib/kubelet/pods/e54ca63f-0568-4b0f-aa00-a726ead780cc/volumes" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.270420 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r"] Mar 12 13:15:52 crc kubenswrapper[4778]: W0312 13:15:52.270344 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6eebe99_8b82_4720_82c5_c940100859ad.slice/crio-ef118b4d4b8450e0cd8d72c97dec0a0efe275f77e1d6ae4557340cc05a492700 WatchSource:0}: Error finding container ef118b4d4b8450e0cd8d72c97dec0a0efe275f77e1d6ae4557340cc05a492700: Status 404 returned error can't find the container with id ef118b4d4b8450e0cd8d72c97dec0a0efe275f77e1d6ae4557340cc05a492700 Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.467241 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" event={"ID":"f6eebe99-8b82-4720-82c5-c940100859ad","Type":"ContainerStarted","Data":"5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb"} Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.467278 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" event={"ID":"f6eebe99-8b82-4720-82c5-c940100859ad","Type":"ContainerStarted","Data":"ef118b4d4b8450e0cd8d72c97dec0a0efe275f77e1d6ae4557340cc05a492700"} Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.468156 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.470507 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" event={"ID":"c8083240-e16c-40da-9f87-9db45cebfafd","Type":"ContainerStarted","Data":"e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68"} Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.471079 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.475170 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.491410 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" podStartSLOduration=3.4913423740000002 podStartE2EDuration="3.491342374s" podCreationTimestamp="2026-03-12 13:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:52.487304711 +0000 UTC m=+370.936000107" watchObservedRunningTime="2026-03-12 13:15:52.491342374 +0000 UTC m=+370.940037780" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.508367 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" podStartSLOduration=3.5083451820000002 podStartE2EDuration="3.508345182s" podCreationTimestamp="2026-03-12 13:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:52.503496104 +0000 UTC m=+370.952191500" watchObservedRunningTime="2026-03-12 13:15:52.508345182 +0000 UTC m=+370.957040578" Mar 12 13:15:52 crc kubenswrapper[4778]: I0312 13:15:52.811264 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:55 crc kubenswrapper[4778]: I0312 13:15:55.547997 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6"] Mar 12 13:15:55 crc kubenswrapper[4778]: I0312 13:15:55.548594 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" podUID="c8083240-e16c-40da-9f87-9db45cebfafd" containerName="controller-manager" containerID="cri-o://e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68" gracePeriod=30 Mar 12 13:15:55 crc kubenswrapper[4778]: I0312 13:15:55.556126 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r"] Mar 12 13:15:55 crc kubenswrapper[4778]: I0312 13:15:55.556355 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" podUID="f6eebe99-8b82-4720-82c5-c940100859ad" containerName="route-controller-manager" containerID="cri-o://5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb" gracePeriod=30 Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.060640 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.097105 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146693 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc82s\" (UniqueName: \"kubernetes.io/projected/f6eebe99-8b82-4720-82c5-c940100859ad-kube-api-access-qc82s\") pod \"f6eebe99-8b82-4720-82c5-c940100859ad\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146785 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-config\") pod \"f6eebe99-8b82-4720-82c5-c940100859ad\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146852 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8083240-e16c-40da-9f87-9db45cebfafd-serving-cert\") pod \"c8083240-e16c-40da-9f87-9db45cebfafd\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146884 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-client-ca\") pod \"f6eebe99-8b82-4720-82c5-c940100859ad\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146898 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6eebe99-8b82-4720-82c5-c940100859ad-serving-cert\") pod \"f6eebe99-8b82-4720-82c5-c940100859ad\" (UID: \"f6eebe99-8b82-4720-82c5-c940100859ad\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146932 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-proxy-ca-bundles\") pod \"c8083240-e16c-40da-9f87-9db45cebfafd\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146953 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-config\") pod \"c8083240-e16c-40da-9f87-9db45cebfafd\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.146997 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-client-ca\") pod \"c8083240-e16c-40da-9f87-9db45cebfafd\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.147019 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jktc4\" (UniqueName: \"kubernetes.io/projected/c8083240-e16c-40da-9f87-9db45cebfafd-kube-api-access-jktc4\") pod \"c8083240-e16c-40da-9f87-9db45cebfafd\" (UID: \"c8083240-e16c-40da-9f87-9db45cebfafd\") " Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.147500 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-client-ca" (OuterVolumeSpecName: "client-ca") pod "f6eebe99-8b82-4720-82c5-c940100859ad" (UID: "f6eebe99-8b82-4720-82c5-c940100859ad"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.147552 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-config" (OuterVolumeSpecName: "config") pod "f6eebe99-8b82-4720-82c5-c940100859ad" (UID: "f6eebe99-8b82-4720-82c5-c940100859ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.147820 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-config" (OuterVolumeSpecName: "config") pod "c8083240-e16c-40da-9f87-9db45cebfafd" (UID: "c8083240-e16c-40da-9f87-9db45cebfafd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.147970 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-client-ca" (OuterVolumeSpecName: "client-ca") pod "c8083240-e16c-40da-9f87-9db45cebfafd" (UID: "c8083240-e16c-40da-9f87-9db45cebfafd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.149336 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c8083240-e16c-40da-9f87-9db45cebfafd" (UID: "c8083240-e16c-40da-9f87-9db45cebfafd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.151602 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.152460 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.152471 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6eebe99-8b82-4720-82c5-c940100859ad-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.152479 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.152491 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8083240-e16c-40da-9f87-9db45cebfafd-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.151784 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8083240-e16c-40da-9f87-9db45cebfafd-kube-api-access-jktc4" (OuterVolumeSpecName: "kube-api-access-jktc4") pod "c8083240-e16c-40da-9f87-9db45cebfafd" (UID: "c8083240-e16c-40da-9f87-9db45cebfafd"). InnerVolumeSpecName "kube-api-access-jktc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.151802 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6eebe99-8b82-4720-82c5-c940100859ad-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f6eebe99-8b82-4720-82c5-c940100859ad" (UID: "f6eebe99-8b82-4720-82c5-c940100859ad"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.151834 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6eebe99-8b82-4720-82c5-c940100859ad-kube-api-access-qc82s" (OuterVolumeSpecName: "kube-api-access-qc82s") pod "f6eebe99-8b82-4720-82c5-c940100859ad" (UID: "f6eebe99-8b82-4720-82c5-c940100859ad"). InnerVolumeSpecName "kube-api-access-qc82s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.151973 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8083240-e16c-40da-9f87-9db45cebfafd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c8083240-e16c-40da-9f87-9db45cebfafd" (UID: "c8083240-e16c-40da-9f87-9db45cebfafd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.254060 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6eebe99-8b82-4720-82c5-c940100859ad-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.254098 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jktc4\" (UniqueName: \"kubernetes.io/projected/c8083240-e16c-40da-9f87-9db45cebfafd-kube-api-access-jktc4\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.254109 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc82s\" (UniqueName: \"kubernetes.io/projected/f6eebe99-8b82-4720-82c5-c940100859ad-kube-api-access-qc82s\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.254117 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8083240-e16c-40da-9f87-9db45cebfafd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.499164 4778 generic.go:334] "Generic (PLEG): container finished" podID="f6eebe99-8b82-4720-82c5-c940100859ad" containerID="5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb" exitCode=0 Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.499268 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" event={"ID":"f6eebe99-8b82-4720-82c5-c940100859ad","Type":"ContainerDied","Data":"5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb"} Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.499284 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.499307 4778 scope.go:117] "RemoveContainer" containerID="5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.499296 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r" event={"ID":"f6eebe99-8b82-4720-82c5-c940100859ad","Type":"ContainerDied","Data":"ef118b4d4b8450e0cd8d72c97dec0a0efe275f77e1d6ae4557340cc05a492700"} Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.502010 4778 generic.go:334] "Generic (PLEG): container finished" podID="c8083240-e16c-40da-9f87-9db45cebfafd" containerID="e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68" exitCode=0 Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.502031 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" event={"ID":"c8083240-e16c-40da-9f87-9db45cebfafd","Type":"ContainerDied","Data":"e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68"} Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.502045 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" event={"ID":"c8083240-e16c-40da-9f87-9db45cebfafd","Type":"ContainerDied","Data":"610f9e9935be3de3cd47d5e81cba88f5a4f3f12d0d389042b1b5255df2fa7476"} Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.502101 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.524908 4778 scope.go:117] "RemoveContainer" containerID="5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb" Mar 12 13:15:56 crc kubenswrapper[4778]: E0312 13:15:56.525507 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb\": container with ID starting with 5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb not found: ID does not exist" containerID="5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.525558 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb"} err="failed to get container status \"5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb\": rpc error: code = NotFound desc = could not find container \"5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb\": container with ID starting with 5e8409c01768716b0eb390d5602b10ae0c1d9381bd08aae1e8d64b1f9635a1eb not found: ID does not exist" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.525593 4778 scope.go:117] "RemoveContainer" containerID="e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.527206 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.534298 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-n4jk6"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.537805 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.541515 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-ggb6r"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.549560 4778 scope.go:117] "RemoveContainer" containerID="e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68" Mar 12 13:15:56 crc kubenswrapper[4778]: E0312 13:15:56.549988 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68\": container with ID starting with e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68 not found: ID does not exist" containerID="e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.550032 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68"} err="failed to get container status \"e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68\": rpc error: code = NotFound desc = could not find container \"e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68\": container with ID starting with e41fa41e2ced748ad127ed8555daf5859b90c9e62957331490bce01447075c68 not found: ID does not exist" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.681573 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f558b8664-wcwww"] Mar 12 13:15:56 crc kubenswrapper[4778]: E0312 13:15:56.682021 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6eebe99-8b82-4720-82c5-c940100859ad" containerName="route-controller-manager" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.682047 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6eebe99-8b82-4720-82c5-c940100859ad" containerName="route-controller-manager" Mar 12 13:15:56 crc kubenswrapper[4778]: E0312 13:15:56.682074 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8083240-e16c-40da-9f87-9db45cebfafd" containerName="controller-manager" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.682091 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8083240-e16c-40da-9f87-9db45cebfafd" containerName="controller-manager" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.682335 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6eebe99-8b82-4720-82c5-c940100859ad" containerName="route-controller-manager" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.682374 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8083240-e16c-40da-9f87-9db45cebfafd" containerName="controller-manager" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.683242 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.687170 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.688939 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.699043 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.699282 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.699478 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.700113 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.700891 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.709622 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.711239 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.715860 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.715912 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.716521 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.716533 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.716879 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.716931 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.718746 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.730418 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f558b8664-wcwww"] Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761414 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-config\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761491 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-config\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761524 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-serving-cert\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761592 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-proxy-ca-bundles\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761618 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f2c27d-f47b-4bcb-81af-749dd8f6d053-serving-cert\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761641 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm445\" (UniqueName: \"kubernetes.io/projected/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-kube-api-access-sm445\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761667 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnhvm\" (UniqueName: \"kubernetes.io/projected/81f2c27d-f47b-4bcb-81af-749dd8f6d053-kube-api-access-vnhvm\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761698 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-client-ca\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.761720 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-client-ca\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862513 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-config\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862610 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-serving-cert\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862690 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-proxy-ca-bundles\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862740 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f2c27d-f47b-4bcb-81af-749dd8f6d053-serving-cert\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862797 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm445\" (UniqueName: \"kubernetes.io/projected/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-kube-api-access-sm445\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862857 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnhvm\" (UniqueName: \"kubernetes.io/projected/81f2c27d-f47b-4bcb-81af-749dd8f6d053-kube-api-access-vnhvm\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862921 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-client-ca\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.862970 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-client-ca\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.863082 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-config\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.864776 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-client-ca\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.864992 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-client-ca\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.865438 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-config\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.865710 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-config\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.866347 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-proxy-ca-bundles\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.868634 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-serving-cert\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.870545 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f2c27d-f47b-4bcb-81af-749dd8f6d053-serving-cert\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.886377 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm445\" (UniqueName: \"kubernetes.io/projected/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-kube-api-access-sm445\") pod \"route-controller-manager-7864ddbcd6-8t87n\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:56 crc kubenswrapper[4778]: I0312 13:15:56.886820 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnhvm\" (UniqueName: \"kubernetes.io/projected/81f2c27d-f47b-4bcb-81af-749dd8f6d053-kube-api-access-vnhvm\") pod \"controller-manager-7f558b8664-wcwww\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:57 crc kubenswrapper[4778]: I0312 13:15:57.033717 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:57 crc kubenswrapper[4778]: I0312 13:15:57.041823 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:57 crc kubenswrapper[4778]: I0312 13:15:57.466003 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f558b8664-wcwww"] Mar 12 13:15:57 crc kubenswrapper[4778]: I0312 13:15:57.511303 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n"] Mar 12 13:15:57 crc kubenswrapper[4778]: W0312 13:15:57.513904 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03ce725f_d022_4fe6_9fd8_d61f4bec2ad6.slice/crio-f2dfa420c929336f07fb954275af02bba9b37d1b1afb3796a9bd8590cd2100e8 WatchSource:0}: Error finding container f2dfa420c929336f07fb954275af02bba9b37d1b1afb3796a9bd8590cd2100e8: Status 404 returned error can't find the container with id f2dfa420c929336f07fb954275af02bba9b37d1b1afb3796a9bd8590cd2100e8 Mar 12 13:15:57 crc kubenswrapper[4778]: I0312 13:15:57.519422 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" event={"ID":"81f2c27d-f47b-4bcb-81af-749dd8f6d053","Type":"ContainerStarted","Data":"6d515fcf6020e53c6189e977403cfe43fcda12d2f5abd6575280ee4b45363384"} Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.264622 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8083240-e16c-40da-9f87-9db45cebfafd" path="/var/lib/kubelet/pods/c8083240-e16c-40da-9f87-9db45cebfafd/volumes" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.265781 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6eebe99-8b82-4720-82c5-c940100859ad" path="/var/lib/kubelet/pods/f6eebe99-8b82-4720-82c5-c940100859ad/volumes" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.526891 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" event={"ID":"81f2c27d-f47b-4bcb-81af-749dd8f6d053","Type":"ContainerStarted","Data":"f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783"} Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.527319 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.528958 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" event={"ID":"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6","Type":"ContainerStarted","Data":"2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe"} Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.529047 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" event={"ID":"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6","Type":"ContainerStarted","Data":"f2dfa420c929336f07fb954275af02bba9b37d1b1afb3796a9bd8590cd2100e8"} Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.529176 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.533391 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.534291 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.549080 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" podStartSLOduration=3.549055172 podStartE2EDuration="3.549055172s" podCreationTimestamp="2026-03-12 13:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:58.545585396 +0000 UTC m=+376.994280802" watchObservedRunningTime="2026-03-12 13:15:58.549055172 +0000 UTC m=+376.997750568" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.557993 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.558089 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:15:58 crc kubenswrapper[4778]: I0312 13:15:58.573252 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" podStartSLOduration=3.573230929 podStartE2EDuration="3.573230929s" podCreationTimestamp="2026-03-12 13:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:15:58.572441195 +0000 UTC m=+377.021136591" watchObservedRunningTime="2026-03-12 13:15:58.573230929 +0000 UTC m=+377.021926325" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.137674 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555356-cdmcz"] Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.138359 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.144717 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.144886 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.147446 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.151506 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555356-cdmcz"] Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.206787 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8btl\" (UniqueName: \"kubernetes.io/projected/c792e81a-8273-49a7-be95-c8c19cd2785b-kube-api-access-v8btl\") pod \"auto-csr-approver-29555356-cdmcz\" (UID: \"c792e81a-8273-49a7-be95-c8c19cd2785b\") " pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.308095 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8btl\" (UniqueName: \"kubernetes.io/projected/c792e81a-8273-49a7-be95-c8c19cd2785b-kube-api-access-v8btl\") pod \"auto-csr-approver-29555356-cdmcz\" (UID: \"c792e81a-8273-49a7-be95-c8c19cd2785b\") " pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.328114 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8btl\" (UniqueName: \"kubernetes.io/projected/c792e81a-8273-49a7-be95-c8c19cd2785b-kube-api-access-v8btl\") pod \"auto-csr-approver-29555356-cdmcz\" (UID: \"c792e81a-8273-49a7-be95-c8c19cd2785b\") " pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.461744 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:00 crc kubenswrapper[4778]: I0312 13:16:00.872550 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555356-cdmcz"] Mar 12 13:16:00 crc kubenswrapper[4778]: W0312 13:16:00.889405 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc792e81a_8273_49a7_be95_c8c19cd2785b.slice/crio-0325104c586ee885bb03aa8d7f2350ccf7dd9664fad4b2303bbd38426cc6f204 WatchSource:0}: Error finding container 0325104c586ee885bb03aa8d7f2350ccf7dd9664fad4b2303bbd38426cc6f204: Status 404 returned error can't find the container with id 0325104c586ee885bb03aa8d7f2350ccf7dd9664fad4b2303bbd38426cc6f204 Mar 12 13:16:01 crc kubenswrapper[4778]: I0312 13:16:01.548388 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" event={"ID":"c792e81a-8273-49a7-be95-c8c19cd2785b","Type":"ContainerStarted","Data":"0325104c586ee885bb03aa8d7f2350ccf7dd9664fad4b2303bbd38426cc6f204"} Mar 12 13:16:03 crc kubenswrapper[4778]: I0312 13:16:03.559158 4778 generic.go:334] "Generic (PLEG): container finished" podID="c792e81a-8273-49a7-be95-c8c19cd2785b" containerID="b6d55e4553c4a90b5714d39c88d9e361c3f3109a89cdbda1980233a5b1fade38" exitCode=0 Mar 12 13:16:03 crc kubenswrapper[4778]: I0312 13:16:03.559219 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" event={"ID":"c792e81a-8273-49a7-be95-c8c19cd2785b","Type":"ContainerDied","Data":"b6d55e4553c4a90b5714d39c88d9e361c3f3109a89cdbda1980233a5b1fade38"} Mar 12 13:16:04 crc kubenswrapper[4778]: I0312 13:16:04.879095 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:05 crc kubenswrapper[4778]: I0312 13:16:05.068035 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8btl\" (UniqueName: \"kubernetes.io/projected/c792e81a-8273-49a7-be95-c8c19cd2785b-kube-api-access-v8btl\") pod \"c792e81a-8273-49a7-be95-c8c19cd2785b\" (UID: \"c792e81a-8273-49a7-be95-c8c19cd2785b\") " Mar 12 13:16:05 crc kubenswrapper[4778]: I0312 13:16:05.072640 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c792e81a-8273-49a7-be95-c8c19cd2785b-kube-api-access-v8btl" (OuterVolumeSpecName: "kube-api-access-v8btl") pod "c792e81a-8273-49a7-be95-c8c19cd2785b" (UID: "c792e81a-8273-49a7-be95-c8c19cd2785b"). InnerVolumeSpecName "kube-api-access-v8btl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:16:05 crc kubenswrapper[4778]: I0312 13:16:05.169578 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8btl\" (UniqueName: \"kubernetes.io/projected/c792e81a-8273-49a7-be95-c8c19cd2785b-kube-api-access-v8btl\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:05 crc kubenswrapper[4778]: I0312 13:16:05.574763 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" event={"ID":"c792e81a-8273-49a7-be95-c8c19cd2785b","Type":"ContainerDied","Data":"0325104c586ee885bb03aa8d7f2350ccf7dd9664fad4b2303bbd38426cc6f204"} Mar 12 13:16:05 crc kubenswrapper[4778]: I0312 13:16:05.574814 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0325104c586ee885bb03aa8d7f2350ccf7dd9664fad4b2303bbd38426cc6f204" Mar 12 13:16:05 crc kubenswrapper[4778]: I0312 13:16:05.574904 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555356-cdmcz" Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.029339 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l8n9b"] Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.031347 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l8n9b" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="registry-server" containerID="cri-o://3686a4e289950327029466c928723a8314f5dcaa797637ff0db63d9aa4aeb5db" gracePeriod=2 Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.225938 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtjz5"] Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.226438 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rtjz5" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="registry-server" containerID="cri-o://3151ddc8cb64182fd7ccd241e4580f2e0243328e43f1e59366f60b980b160490" gracePeriod=2 Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.664662 4778 generic.go:334] "Generic (PLEG): container finished" podID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerID="3151ddc8cb64182fd7ccd241e4580f2e0243328e43f1e59366f60b980b160490" exitCode=0 Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.664754 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtjz5" event={"ID":"b9bef112-9bef-4ce2-abd8-054b4d671658","Type":"ContainerDied","Data":"3151ddc8cb64182fd7ccd241e4580f2e0243328e43f1e59366f60b980b160490"} Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.670468 4778 generic.go:334] "Generic (PLEG): container finished" podID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerID="3686a4e289950327029466c928723a8314f5dcaa797637ff0db63d9aa4aeb5db" exitCode=0 Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.670513 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l8n9b" event={"ID":"c27afe2a-3402-49f9-b985-45fe67e40d22","Type":"ContainerDied","Data":"3686a4e289950327029466c928723a8314f5dcaa797637ff0db63d9aa4aeb5db"} Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.816757 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.944135 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjm7s\" (UniqueName: \"kubernetes.io/projected/b9bef112-9bef-4ce2-abd8-054b4d671658-kube-api-access-gjm7s\") pod \"b9bef112-9bef-4ce2-abd8-054b4d671658\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.944210 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-utilities\") pod \"b9bef112-9bef-4ce2-abd8-054b4d671658\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.944277 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-catalog-content\") pod \"b9bef112-9bef-4ce2-abd8-054b4d671658\" (UID: \"b9bef112-9bef-4ce2-abd8-054b4d671658\") " Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.945583 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-utilities" (OuterVolumeSpecName: "utilities") pod "b9bef112-9bef-4ce2-abd8-054b4d671658" (UID: "b9bef112-9bef-4ce2-abd8-054b4d671658"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:16:19 crc kubenswrapper[4778]: I0312 13:16:19.978382 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9bef112-9bef-4ce2-abd8-054b4d671658-kube-api-access-gjm7s" (OuterVolumeSpecName: "kube-api-access-gjm7s") pod "b9bef112-9bef-4ce2-abd8-054b4d671658" (UID: "b9bef112-9bef-4ce2-abd8-054b4d671658"). InnerVolumeSpecName "kube-api-access-gjm7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.002046 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9bef112-9bef-4ce2-abd8-054b4d671658" (UID: "b9bef112-9bef-4ce2-abd8-054b4d671658"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.045540 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.045574 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjm7s\" (UniqueName: \"kubernetes.io/projected/b9bef112-9bef-4ce2-abd8-054b4d671658-kube-api-access-gjm7s\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.045588 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9bef112-9bef-4ce2-abd8-054b4d671658-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.174599 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.349419 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxz76\" (UniqueName: \"kubernetes.io/projected/c27afe2a-3402-49f9-b985-45fe67e40d22-kube-api-access-dxz76\") pod \"c27afe2a-3402-49f9-b985-45fe67e40d22\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.349546 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-utilities\") pod \"c27afe2a-3402-49f9-b985-45fe67e40d22\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.349626 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-catalog-content\") pod \"c27afe2a-3402-49f9-b985-45fe67e40d22\" (UID: \"c27afe2a-3402-49f9-b985-45fe67e40d22\") " Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.350362 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-utilities" (OuterVolumeSpecName: "utilities") pod "c27afe2a-3402-49f9-b985-45fe67e40d22" (UID: "c27afe2a-3402-49f9-b985-45fe67e40d22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.352600 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c27afe2a-3402-49f9-b985-45fe67e40d22-kube-api-access-dxz76" (OuterVolumeSpecName: "kube-api-access-dxz76") pod "c27afe2a-3402-49f9-b985-45fe67e40d22" (UID: "c27afe2a-3402-49f9-b985-45fe67e40d22"). InnerVolumeSpecName "kube-api-access-dxz76". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.412923 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c27afe2a-3402-49f9-b985-45fe67e40d22" (UID: "c27afe2a-3402-49f9-b985-45fe67e40d22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.451210 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxz76\" (UniqueName: \"kubernetes.io/projected/c27afe2a-3402-49f9-b985-45fe67e40d22-kube-api-access-dxz76\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.451254 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.451268 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27afe2a-3402-49f9-b985-45fe67e40d22-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.688498 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtjz5" event={"ID":"b9bef112-9bef-4ce2-abd8-054b4d671658","Type":"ContainerDied","Data":"823af2a7e3b6063a4f30d49b66161c625efcb36bf067f9d539324e41889ea011"} Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.688553 4778 scope.go:117] "RemoveContainer" containerID="3151ddc8cb64182fd7ccd241e4580f2e0243328e43f1e59366f60b980b160490" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.688705 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtjz5" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.697080 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l8n9b" event={"ID":"c27afe2a-3402-49f9-b985-45fe67e40d22","Type":"ContainerDied","Data":"f3e464dc52992fdb0f0b53c632c09c98afcb767da1a2f76ffc34b25c53dcb6a3"} Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.697207 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l8n9b" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.711861 4778 scope.go:117] "RemoveContainer" containerID="44212f253b9d8de159bf039fe64dd134b5f7beb71943da6aab7d4efc080466b3" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.718294 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtjz5"] Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.723487 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtjz5"] Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.724105 4778 scope.go:117] "RemoveContainer" containerID="3ee91beb1526d7d2135a66716b66577b22ca3756c6f18236717330ab9060a779" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.736230 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l8n9b"] Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.740014 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l8n9b"] Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.743927 4778 scope.go:117] "RemoveContainer" containerID="3686a4e289950327029466c928723a8314f5dcaa797637ff0db63d9aa4aeb5db" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.767964 4778 scope.go:117] "RemoveContainer" containerID="517c2af638efb950196e9ef53f4578b28c6c02cc9d241b33a72ede0303af599d" Mar 12 13:16:20 crc kubenswrapper[4778]: I0312 13:16:20.781068 4778 scope.go:117] "RemoveContainer" containerID="beac9341cf9caf9b2899c0d3555998167e4413386821c255145cfe1b113c1402" Mar 12 13:16:21 crc kubenswrapper[4778]: I0312 13:16:21.627057 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sjk9p"] Mar 12 13:16:21 crc kubenswrapper[4778]: I0312 13:16:21.628016 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sjk9p" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="registry-server" containerID="cri-o://7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06" gracePeriod=2 Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.262666 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" path="/var/lib/kubelet/pods/b9bef112-9bef-4ce2-abd8-054b4d671658/volumes" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.264152 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" path="/var/lib/kubelet/pods/c27afe2a-3402-49f9-b985-45fe67e40d22/volumes" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.670442 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.711677 4778 generic.go:334] "Generic (PLEG): container finished" podID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerID="7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06" exitCode=0 Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.711720 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjk9p" event={"ID":"3b3fb69e-dd4f-4787-a207-4fe25106f9e7","Type":"ContainerDied","Data":"7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06"} Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.711732 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjk9p" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.711750 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjk9p" event={"ID":"3b3fb69e-dd4f-4787-a207-4fe25106f9e7","Type":"ContainerDied","Data":"54d14e24e2014de0b1846a5aa684b84b3bf2783c8e0d47fb26e64cb9f10b0a8d"} Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.711771 4778 scope.go:117] "RemoveContainer" containerID="7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.745391 4778 scope.go:117] "RemoveContainer" containerID="fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.775775 4778 scope.go:117] "RemoveContainer" containerID="abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.789408 4778 scope.go:117] "RemoveContainer" containerID="7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06" Mar 12 13:16:22 crc kubenswrapper[4778]: E0312 13:16:22.791231 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06\": container with ID starting with 7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06 not found: ID does not exist" containerID="7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.791283 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06"} err="failed to get container status \"7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06\": rpc error: code = NotFound desc = could not find container \"7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06\": container with ID starting with 7a538b433370f97911f22dbc738a9c42cbd5e516b7acdb71010394ade11cee06 not found: ID does not exist" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.791315 4778 scope.go:117] "RemoveContainer" containerID="fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268" Mar 12 13:16:22 crc kubenswrapper[4778]: E0312 13:16:22.791702 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268\": container with ID starting with fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268 not found: ID does not exist" containerID="fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.791735 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268"} err="failed to get container status \"fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268\": rpc error: code = NotFound desc = could not find container \"fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268\": container with ID starting with fa00faf2580a0c0e9d72ea15f4cf1840ea1708c190198951e6018c60afdde268 not found: ID does not exist" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.791757 4778 scope.go:117] "RemoveContainer" containerID="abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961" Mar 12 13:16:22 crc kubenswrapper[4778]: E0312 13:16:22.792149 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961\": container with ID starting with abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961 not found: ID does not exist" containerID="abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.792227 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961"} err="failed to get container status \"abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961\": rpc error: code = NotFound desc = could not find container \"abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961\": container with ID starting with abeebebb9ab695d88020f3373974a8763b6d3a7633ca84c98e6d48516351c961 not found: ID does not exist" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.800597 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq7jg\" (UniqueName: \"kubernetes.io/projected/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-kube-api-access-dq7jg\") pod \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.800671 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-utilities\") pod \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.800719 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-catalog-content\") pod \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\" (UID: \"3b3fb69e-dd4f-4787-a207-4fe25106f9e7\") " Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.801703 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-utilities" (OuterVolumeSpecName: "utilities") pod "3b3fb69e-dd4f-4787-a207-4fe25106f9e7" (UID: "3b3fb69e-dd4f-4787-a207-4fe25106f9e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.809379 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-kube-api-access-dq7jg" (OuterVolumeSpecName: "kube-api-access-dq7jg") pod "3b3fb69e-dd4f-4787-a207-4fe25106f9e7" (UID: "3b3fb69e-dd4f-4787-a207-4fe25106f9e7"). InnerVolumeSpecName "kube-api-access-dq7jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.850304 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b3fb69e-dd4f-4787-a207-4fe25106f9e7" (UID: "3b3fb69e-dd4f-4787-a207-4fe25106f9e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.901584 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.901618 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:22 crc kubenswrapper[4778]: I0312 13:16:22.901632 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq7jg\" (UniqueName: \"kubernetes.io/projected/3b3fb69e-dd4f-4787-a207-4fe25106f9e7-kube-api-access-dq7jg\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:23 crc kubenswrapper[4778]: I0312 13:16:23.051445 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sjk9p"] Mar 12 13:16:23 crc kubenswrapper[4778]: I0312 13:16:23.057082 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sjk9p"] Mar 12 13:16:24 crc kubenswrapper[4778]: I0312 13:16:24.260448 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" path="/var/lib/kubelet/pods/3b3fb69e-dd4f-4787-a207-4fe25106f9e7/volumes" Mar 12 13:16:28 crc kubenswrapper[4778]: I0312 13:16:28.557811 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:16:28 crc kubenswrapper[4778]: I0312 13:16:28.558104 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:16:33 crc kubenswrapper[4778]: I0312 13:16:33.896406 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7f558b8664-wcwww"] Mar 12 13:16:33 crc kubenswrapper[4778]: I0312 13:16:33.896915 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" podUID="81f2c27d-f47b-4bcb-81af-749dd8f6d053" containerName="controller-manager" containerID="cri-o://f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783" gracePeriod=30 Mar 12 13:16:33 crc kubenswrapper[4778]: I0312 13:16:33.911871 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n"] Mar 12 13:16:33 crc kubenswrapper[4778]: I0312 13:16:33.912115 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" podUID="03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" containerName="route-controller-manager" containerID="cri-o://2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe" gracePeriod=30 Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.398364 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.500174 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.548037 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-client-ca\") pod \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.548109 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-config\") pod \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.548195 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-serving-cert\") pod \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.548250 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm445\" (UniqueName: \"kubernetes.io/projected/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-kube-api-access-sm445\") pod \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\" (UID: \"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.548887 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-client-ca" (OuterVolumeSpecName: "client-ca") pod "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" (UID: "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.548945 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-config" (OuterVolumeSpecName: "config") pod "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" (UID: "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.553353 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" (UID: "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.553588 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-kube-api-access-sm445" (OuterVolumeSpecName: "kube-api-access-sm445") pod "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" (UID: "03ce725f-d022-4fe6-9fd8-d61f4bec2ad6"). InnerVolumeSpecName "kube-api-access-sm445". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.649479 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-config\") pod \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.649579 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f2c27d-f47b-4bcb-81af-749dd8f6d053-serving-cert\") pod \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.649605 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnhvm\" (UniqueName: \"kubernetes.io/projected/81f2c27d-f47b-4bcb-81af-749dd8f6d053-kube-api-access-vnhvm\") pod \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.649645 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-client-ca\") pod \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.649766 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-proxy-ca-bundles\") pod \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\" (UID: \"81f2c27d-f47b-4bcb-81af-749dd8f6d053\") " Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650017 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650034 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm445\" (UniqueName: \"kubernetes.io/projected/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-kube-api-access-sm445\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650047 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650060 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650419 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "81f2c27d-f47b-4bcb-81af-749dd8f6d053" (UID: "81f2c27d-f47b-4bcb-81af-749dd8f6d053"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650453 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-config" (OuterVolumeSpecName: "config") pod "81f2c27d-f47b-4bcb-81af-749dd8f6d053" (UID: "81f2c27d-f47b-4bcb-81af-749dd8f6d053"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.650544 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-client-ca" (OuterVolumeSpecName: "client-ca") pod "81f2c27d-f47b-4bcb-81af-749dd8f6d053" (UID: "81f2c27d-f47b-4bcb-81af-749dd8f6d053"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.652703 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f2c27d-f47b-4bcb-81af-749dd8f6d053-kube-api-access-vnhvm" (OuterVolumeSpecName: "kube-api-access-vnhvm") pod "81f2c27d-f47b-4bcb-81af-749dd8f6d053" (UID: "81f2c27d-f47b-4bcb-81af-749dd8f6d053"). InnerVolumeSpecName "kube-api-access-vnhvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.653796 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f2c27d-f47b-4bcb-81af-749dd8f6d053-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "81f2c27d-f47b-4bcb-81af-749dd8f6d053" (UID: "81f2c27d-f47b-4bcb-81af-749dd8f6d053"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.750810 4778 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.750853 4778 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f2c27d-f47b-4bcb-81af-749dd8f6d053-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.750868 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnhvm\" (UniqueName: \"kubernetes.io/projected/81f2c27d-f47b-4bcb-81af-749dd8f6d053-kube-api-access-vnhvm\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.750882 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.750893 4778 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81f2c27d-f47b-4bcb-81af-749dd8f6d053-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.808395 4778 generic.go:334] "Generic (PLEG): container finished" podID="03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" containerID="2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe" exitCode=0 Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.808468 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" event={"ID":"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6","Type":"ContainerDied","Data":"2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe"} Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.808494 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" event={"ID":"03ce725f-d022-4fe6-9fd8-d61f4bec2ad6","Type":"ContainerDied","Data":"f2dfa420c929336f07fb954275af02bba9b37d1b1afb3796a9bd8590cd2100e8"} Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.808512 4778 scope.go:117] "RemoveContainer" containerID="2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.808531 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.815270 4778 generic.go:334] "Generic (PLEG): container finished" podID="81f2c27d-f47b-4bcb-81af-749dd8f6d053" containerID="f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783" exitCode=0 Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.815332 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" event={"ID":"81f2c27d-f47b-4bcb-81af-749dd8f6d053","Type":"ContainerDied","Data":"f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783"} Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.815371 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" event={"ID":"81f2c27d-f47b-4bcb-81af-749dd8f6d053","Type":"ContainerDied","Data":"6d515fcf6020e53c6189e977403cfe43fcda12d2f5abd6575280ee4b45363384"} Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.815446 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f558b8664-wcwww" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.837356 4778 scope.go:117] "RemoveContainer" containerID="2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe" Mar 12 13:16:34 crc kubenswrapper[4778]: E0312 13:16:34.838048 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe\": container with ID starting with 2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe not found: ID does not exist" containerID="2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.838082 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe"} err="failed to get container status \"2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe\": rpc error: code = NotFound desc = could not find container \"2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe\": container with ID starting with 2cab64649829ce76f15124685e98299c4b83afa38015be406f773626bc1243fe not found: ID does not exist" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.838104 4778 scope.go:117] "RemoveContainer" containerID="f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.864911 4778 scope.go:117] "RemoveContainer" containerID="f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783" Mar 12 13:16:34 crc kubenswrapper[4778]: E0312 13:16:34.865470 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783\": container with ID starting with f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783 not found: ID does not exist" containerID="f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.865554 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783"} err="failed to get container status \"f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783\": rpc error: code = NotFound desc = could not find container \"f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783\": container with ID starting with f2e5031b27b99dfbb22f5b0690baf59159234c18ae5db2065cb1c9e1a7bfc783 not found: ID does not exist" Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.866554 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7f558b8664-wcwww"] Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.872061 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7f558b8664-wcwww"] Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.876253 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n"] Mar 12 13:16:34 crc kubenswrapper[4778]: I0312 13:16:34.880049 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7864ddbcd6-8t87n"] Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.698527 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6"] Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699294 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="extract-utilities" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699318 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="extract-utilities" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699343 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="extract-utilities" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699355 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="extract-utilities" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699372 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="extract-content" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699383 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="extract-content" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699398 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f2c27d-f47b-4bcb-81af-749dd8f6d053" containerName="controller-manager" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699410 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f2c27d-f47b-4bcb-81af-749dd8f6d053" containerName="controller-manager" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699429 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="extract-content" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699439 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="extract-content" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699457 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c792e81a-8273-49a7-be95-c8c19cd2785b" containerName="oc" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699467 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c792e81a-8273-49a7-be95-c8c19cd2785b" containerName="oc" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699482 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" containerName="route-controller-manager" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699493 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" containerName="route-controller-manager" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699509 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699519 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699535 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699545 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699566 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699579 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699601 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="extract-content" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699612 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="extract-content" Mar 12 13:16:35 crc kubenswrapper[4778]: E0312 13:16:35.699626 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="extract-utilities" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699637 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="extract-utilities" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699803 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f2c27d-f47b-4bcb-81af-749dd8f6d053" containerName="controller-manager" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699823 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c792e81a-8273-49a7-be95-c8c19cd2785b" containerName="oc" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699839 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b3fb69e-dd4f-4787-a207-4fe25106f9e7" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699857 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" containerName="route-controller-manager" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699870 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c27afe2a-3402-49f9-b985-45fe67e40d22" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.699884 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9bef112-9bef-4ce2-abd8-054b4d671658" containerName="registry-server" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.700492 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.705799 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.705979 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.706270 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.706397 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.706447 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.706447 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.709810 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-888fq"] Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.710483 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.713140 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.713875 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.713912 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.713937 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.714259 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.714847 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.718677 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6"] Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.723079 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.727299 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-888fq"] Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764353 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-client-ca\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764400 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c3b72a7-abe5-4c24-b14b-04ae34b28816-serving-cert\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764425 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2rdp\" (UniqueName: \"kubernetes.io/projected/5c3b72a7-abe5-4c24-b14b-04ae34b28816-kube-api-access-n2rdp\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764443 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e76627c-6dee-4814-bc83-672b4350b105-serving-cert\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764522 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-config\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764585 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-proxy-ca-bundles\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764621 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9gbx\" (UniqueName: \"kubernetes.io/projected/6e76627c-6dee-4814-bc83-672b4350b105-kube-api-access-v9gbx\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764643 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c3b72a7-abe5-4c24-b14b-04ae34b28816-client-ca\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.764668 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c3b72a7-abe5-4c24-b14b-04ae34b28816-config\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.865880 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-proxy-ca-bundles\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.865948 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9gbx\" (UniqueName: \"kubernetes.io/projected/6e76627c-6dee-4814-bc83-672b4350b105-kube-api-access-v9gbx\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.865987 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c3b72a7-abe5-4c24-b14b-04ae34b28816-client-ca\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.866018 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c3b72a7-abe5-4c24-b14b-04ae34b28816-config\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.866076 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-client-ca\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.866111 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c3b72a7-abe5-4c24-b14b-04ae34b28816-serving-cert\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.866141 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2rdp\" (UniqueName: \"kubernetes.io/projected/5c3b72a7-abe5-4c24-b14b-04ae34b28816-kube-api-access-n2rdp\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.866163 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e76627c-6dee-4814-bc83-672b4350b105-serving-cert\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.866251 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-config\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.867235 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-client-ca\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.867235 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-proxy-ca-bundles\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.868114 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c3b72a7-abe5-4c24-b14b-04ae34b28816-client-ca\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.868286 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e76627c-6dee-4814-bc83-672b4350b105-config\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.869427 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c3b72a7-abe5-4c24-b14b-04ae34b28816-config\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.872361 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e76627c-6dee-4814-bc83-672b4350b105-serving-cert\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.873377 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c3b72a7-abe5-4c24-b14b-04ae34b28816-serving-cert\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.883298 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9gbx\" (UniqueName: \"kubernetes.io/projected/6e76627c-6dee-4814-bc83-672b4350b105-kube-api-access-v9gbx\") pod \"controller-manager-f8c4b6bf8-888fq\" (UID: \"6e76627c-6dee-4814-bc83-672b4350b105\") " pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:35 crc kubenswrapper[4778]: I0312 13:16:35.885956 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2rdp\" (UniqueName: \"kubernetes.io/projected/5c3b72a7-abe5-4c24-b14b-04ae34b28816-kube-api-access-n2rdp\") pod \"route-controller-manager-6d496c4846-wmjt6\" (UID: \"5c3b72a7-abe5-4c24-b14b-04ae34b28816\") " pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.032175 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.051630 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.260267 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ce725f-d022-4fe6-9fd8-d61f4bec2ad6" path="/var/lib/kubelet/pods/03ce725f-d022-4fe6-9fd8-d61f4bec2ad6/volumes" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.261071 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f2c27d-f47b-4bcb-81af-749dd8f6d053" path="/var/lib/kubelet/pods/81f2c27d-f47b-4bcb-81af-749dd8f6d053/volumes" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.464075 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6"] Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.522634 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8c4b6bf8-888fq"] Mar 12 13:16:36 crc kubenswrapper[4778]: W0312 13:16:36.526686 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e76627c_6dee_4814_bc83_672b4350b105.slice/crio-eb191a09f1ca4d1870d02902ac25b6f70f145d331329db420860af9c43f6f227 WatchSource:0}: Error finding container eb191a09f1ca4d1870d02902ac25b6f70f145d331329db420860af9c43f6f227: Status 404 returned error can't find the container with id eb191a09f1ca4d1870d02902ac25b6f70f145d331329db420860af9c43f6f227 Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.831579 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" event={"ID":"6e76627c-6dee-4814-bc83-672b4350b105","Type":"ContainerStarted","Data":"bc6fea73d079c9b176968cc59a0eb92caf386273ff3cf70afe4c4b8e1092c6e2"} Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.831880 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" event={"ID":"6e76627c-6dee-4814-bc83-672b4350b105","Type":"ContainerStarted","Data":"eb191a09f1ca4d1870d02902ac25b6f70f145d331329db420860af9c43f6f227"} Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.831895 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.834121 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" event={"ID":"5c3b72a7-abe5-4c24-b14b-04ae34b28816","Type":"ContainerStarted","Data":"105ee0ee9503b4985a14e9c66aefc21057c6fad2e578ad711e30cd2d47d5bc36"} Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.834174 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" event={"ID":"5c3b72a7-abe5-4c24-b14b-04ae34b28816","Type":"ContainerStarted","Data":"7af6ed9ee3cdaae266963c648e0a3b0d6683dc449c0a4b975b488b7e3ca35367"} Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.834246 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.844724 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.855489 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f8c4b6bf8-888fq" podStartSLOduration=3.855470967 podStartE2EDuration="3.855470967s" podCreationTimestamp="2026-03-12 13:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:16:36.855152917 +0000 UTC m=+415.303848323" watchObservedRunningTime="2026-03-12 13:16:36.855470967 +0000 UTC m=+415.304166363" Mar 12 13:16:36 crc kubenswrapper[4778]: I0312 13:16:36.918219 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" podStartSLOduration=2.918177365 podStartE2EDuration="2.918177365s" podCreationTimestamp="2026-03-12 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:16:36.90298566 +0000 UTC m=+415.351681056" watchObservedRunningTime="2026-03-12 13:16:36.918177365 +0000 UTC m=+415.366872761" Mar 12 13:16:37 crc kubenswrapper[4778]: I0312 13:16:37.176507 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d496c4846-wmjt6" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.402757 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rt8dj"] Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.403760 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.424195 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rt8dj"] Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600262 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600313 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/279e578a-c571-4915-994d-588cf930abe6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600421 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/279e578a-c571-4915-994d-588cf930abe6-registry-certificates\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600520 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk88l\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-kube-api-access-nk88l\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600576 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/279e578a-c571-4915-994d-588cf930abe6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600677 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-bound-sa-token\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600798 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279e578a-c571-4915-994d-588cf930abe6-trusted-ca\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.600901 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-registry-tls\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.622434 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702385 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/279e578a-c571-4915-994d-588cf930abe6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702425 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-bound-sa-token\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702448 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279e578a-c571-4915-994d-588cf930abe6-trusted-ca\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702473 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-registry-tls\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702505 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/279e578a-c571-4915-994d-588cf930abe6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702533 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/279e578a-c571-4915-994d-588cf930abe6-registry-certificates\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.702552 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk88l\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-kube-api-access-nk88l\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.703197 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/279e578a-c571-4915-994d-588cf930abe6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.703611 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279e578a-c571-4915-994d-588cf930abe6-trusted-ca\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.703798 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/279e578a-c571-4915-994d-588cf930abe6-registry-certificates\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.708774 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-registry-tls\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.719886 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/279e578a-c571-4915-994d-588cf930abe6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.720911 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk88l\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-kube-api-access-nk88l\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:38 crc kubenswrapper[4778]: I0312 13:16:38.721492 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/279e578a-c571-4915-994d-588cf930abe6-bound-sa-token\") pod \"image-registry-66df7c8f76-rt8dj\" (UID: \"279e578a-c571-4915-994d-588cf930abe6\") " pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:39 crc kubenswrapper[4778]: I0312 13:16:39.051520 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:39 crc kubenswrapper[4778]: I0312 13:16:39.637818 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rt8dj"] Mar 12 13:16:39 crc kubenswrapper[4778]: W0312 13:16:39.644591 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod279e578a_c571_4915_994d_588cf930abe6.slice/crio-1ee4a31ad528857f01145fbf2df7d6e11f61b0f976cb3844e2d01a2e372d66c3 WatchSource:0}: Error finding container 1ee4a31ad528857f01145fbf2df7d6e11f61b0f976cb3844e2d01a2e372d66c3: Status 404 returned error can't find the container with id 1ee4a31ad528857f01145fbf2df7d6e11f61b0f976cb3844e2d01a2e372d66c3 Mar 12 13:16:39 crc kubenswrapper[4778]: I0312 13:16:39.850639 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" event={"ID":"279e578a-c571-4915-994d-588cf930abe6","Type":"ContainerStarted","Data":"4e507aaa9f1e3645b83dbe1661a940ee2905ea4f684ba9048e16d660a87d550b"} Mar 12 13:16:39 crc kubenswrapper[4778]: I0312 13:16:39.850686 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" event={"ID":"279e578a-c571-4915-994d-588cf930abe6","Type":"ContainerStarted","Data":"1ee4a31ad528857f01145fbf2df7d6e11f61b0f976cb3844e2d01a2e372d66c3"} Mar 12 13:16:39 crc kubenswrapper[4778]: I0312 13:16:39.851385 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:39 crc kubenswrapper[4778]: I0312 13:16:39.872246 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" podStartSLOduration=1.872223076 podStartE2EDuration="1.872223076s" podCreationTimestamp="2026-03-12 13:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:16:39.868028278 +0000 UTC m=+418.316723684" watchObservedRunningTime="2026-03-12 13:16:39.872223076 +0000 UTC m=+418.320918472" Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.557600 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.558289 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.558364 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.559133 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dcabd48eda797c052967d086d455193bf30a1f05151385a52352d733c58148f7"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.559237 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://dcabd48eda797c052967d086d455193bf30a1f05151385a52352d733c58148f7" gracePeriod=600 Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.969719 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="dcabd48eda797c052967d086d455193bf30a1f05151385a52352d733c58148f7" exitCode=0 Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.970004 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"dcabd48eda797c052967d086d455193bf30a1f05151385a52352d733c58148f7"} Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.970031 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"e50690e6aff1fd408e6201d1eee1240e692ce04bc21873dbbe85a5f2d638d704"} Mar 12 13:16:58 crc kubenswrapper[4778]: I0312 13:16:58.970046 4778 scope.go:117] "RemoveContainer" containerID="14daba92184fca91c6930d5b3e821f88408e0fd40a7793f2d70f82df7c9444ce" Mar 12 13:16:59 crc kubenswrapper[4778]: I0312 13:16:59.059566 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rt8dj" Mar 12 13:16:59 crc kubenswrapper[4778]: I0312 13:16:59.125151 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxrx4"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.122024 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qx9d8"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.122680 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qx9d8" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="registry-server" containerID="cri-o://13189da41e0fb30fa7cca9718222038a2b578d40c4f21c5e350b74e753b85587" gracePeriod=30 Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.128427 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-khr6h"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.128685 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-khr6h" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="registry-server" containerID="cri-o://b352e6584b478e7228a408cc5d6c8b18473e75a0de7be819c32ae9b98a707a4e" gracePeriod=30 Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.133283 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2wqm5"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.133511 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" podUID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" containerName="marketplace-operator" containerID="cri-o://013c13acbd136a9ae3c6c39b9470a59aa4ab705637939d6af761af9e92e81b9c" gracePeriod=30 Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.144668 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hvmk8"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.145346 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.160675 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xksl"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.160997 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8xksl" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="registry-server" containerID="cri-o://1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592" gracePeriod=30 Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.172222 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hvmk8"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.182327 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5s5vs"] Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.182616 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5s5vs" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="registry-server" containerID="cri-o://06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd" gracePeriod=30 Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.290837 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b062c23-5acd-430d-aa6c-24b48a725594-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.290911 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2m4b\" (UniqueName: \"kubernetes.io/projected/3b062c23-5acd-430d-aa6c-24b48a725594-kube-api-access-m2m4b\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.290945 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b062c23-5acd-430d-aa6c-24b48a725594-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.392472 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b062c23-5acd-430d-aa6c-24b48a725594-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.392854 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2m4b\" (UniqueName: \"kubernetes.io/projected/3b062c23-5acd-430d-aa6c-24b48a725594-kube-api-access-m2m4b\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.393231 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b062c23-5acd-430d-aa6c-24b48a725594-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.393636 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b062c23-5acd-430d-aa6c-24b48a725594-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.408340 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b062c23-5acd-430d-aa6c-24b48a725594-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.412890 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2m4b\" (UniqueName: \"kubernetes.io/projected/3b062c23-5acd-430d-aa6c-24b48a725594-kube-api-access-m2m4b\") pod \"marketplace-operator-79b997595-hvmk8\" (UID: \"3b062c23-5acd-430d-aa6c-24b48a725594\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.481388 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.701388 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:17:01 crc kubenswrapper[4778]: I0312 13:17:01.781323 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.806013 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-utilities\") pod \"de4557b4-7957-47a0-8c42-845be1fa0f32\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.806116 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-catalog-content\") pod \"de4557b4-7957-47a0-8c42-845be1fa0f32\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.806215 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kch8z\" (UniqueName: \"kubernetes.io/projected/de4557b4-7957-47a0-8c42-845be1fa0f32-kube-api-access-kch8z\") pod \"de4557b4-7957-47a0-8c42-845be1fa0f32\" (UID: \"de4557b4-7957-47a0-8c42-845be1fa0f32\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.807380 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-utilities" (OuterVolumeSpecName: "utilities") pod "de4557b4-7957-47a0-8c42-845be1fa0f32" (UID: "de4557b4-7957-47a0-8c42-845be1fa0f32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.814389 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4557b4-7957-47a0-8c42-845be1fa0f32-kube-api-access-kch8z" (OuterVolumeSpecName: "kube-api-access-kch8z") pod "de4557b4-7957-47a0-8c42-845be1fa0f32" (UID: "de4557b4-7957-47a0-8c42-845be1fa0f32"). InnerVolumeSpecName "kube-api-access-kch8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.835539 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de4557b4-7957-47a0-8c42-845be1fa0f32" (UID: "de4557b4-7957-47a0-8c42-845be1fa0f32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.907900 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-utilities\") pod \"f438f2a3-60c0-4554-a49b-030545f8139c\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.908071 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-catalog-content\") pod \"f438f2a3-60c0-4554-a49b-030545f8139c\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.908826 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-utilities" (OuterVolumeSpecName: "utilities") pod "f438f2a3-60c0-4554-a49b-030545f8139c" (UID: "f438f2a3-60c0-4554-a49b-030545f8139c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.912292 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f438f2a3-60c0-4554-a49b-030545f8139c-kube-api-access-mpfz2" (OuterVolumeSpecName: "kube-api-access-mpfz2") pod "f438f2a3-60c0-4554-a49b-030545f8139c" (UID: "f438f2a3-60c0-4554-a49b-030545f8139c"). InnerVolumeSpecName "kube-api-access-mpfz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.915808 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hvmk8"] Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.908100 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpfz2\" (UniqueName: \"kubernetes.io/projected/f438f2a3-60c0-4554-a49b-030545f8139c-kube-api-access-mpfz2\") pod \"f438f2a3-60c0-4554-a49b-030545f8139c\" (UID: \"f438f2a3-60c0-4554-a49b-030545f8139c\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.917720 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.917735 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4557b4-7957-47a0-8c42-845be1fa0f32-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.917745 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpfz2\" (UniqueName: \"kubernetes.io/projected/f438f2a3-60c0-4554-a49b-030545f8139c-kube-api-access-mpfz2\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.917754 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kch8z\" (UniqueName: \"kubernetes.io/projected/de4557b4-7957-47a0-8c42-845be1fa0f32-kube-api-access-kch8z\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.917762 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.998676 4778 generic.go:334] "Generic (PLEG): container finished" podID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerID="1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592" exitCode=0 Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.998748 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xksl" event={"ID":"de4557b4-7957-47a0-8c42-845be1fa0f32","Type":"ContainerDied","Data":"1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.998780 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xksl" event={"ID":"de4557b4-7957-47a0-8c42-845be1fa0f32","Type":"ContainerDied","Data":"775a67dbf14a4aa00ee320f14ee688f2689c34e66ee23b796f0166af1618f55f"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.998800 4778 scope.go:117] "RemoveContainer" containerID="1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:01.998923 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xksl" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.010464 4778 generic.go:334] "Generic (PLEG): container finished" podID="f438f2a3-60c0-4554-a49b-030545f8139c" containerID="06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd" exitCode=0 Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.010535 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerDied","Data":"06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.010564 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s5vs" event={"ID":"f438f2a3-60c0-4554-a49b-030545f8139c","Type":"ContainerDied","Data":"c5e7e785f566d6c012fb07b0778c4b6c15691ef04836f8607417e605e9c6feb5"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.010564 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s5vs" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.013802 4778 generic.go:334] "Generic (PLEG): container finished" podID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerID="b352e6584b478e7228a408cc5d6c8b18473e75a0de7be819c32ae9b98a707a4e" exitCode=0 Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.013925 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khr6h" event={"ID":"1d185732-cd6b-44c6-b4db-ee9ade00c683","Type":"ContainerDied","Data":"b352e6584b478e7228a408cc5d6c8b18473e75a0de7be819c32ae9b98a707a4e"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.015230 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" event={"ID":"3b062c23-5acd-430d-aa6c-24b48a725594","Type":"ContainerStarted","Data":"fbcdce17a0d9bdb5efce73cec35011a42fcb30d115a6cbf1a834fe9c4ac95c31"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.017500 4778 generic.go:334] "Generic (PLEG): container finished" podID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" containerID="013c13acbd136a9ae3c6c39b9470a59aa4ab705637939d6af761af9e92e81b9c" exitCode=0 Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.017583 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" event={"ID":"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d","Type":"ContainerDied","Data":"013c13acbd136a9ae3c6c39b9470a59aa4ab705637939d6af761af9e92e81b9c"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.023327 4778 generic.go:334] "Generic (PLEG): container finished" podID="651601bd-18fe-4ca1-9c61-481ca568d022" containerID="13189da41e0fb30fa7cca9718222038a2b578d40c4f21c5e350b74e753b85587" exitCode=0 Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.023532 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qx9d8" event={"ID":"651601bd-18fe-4ca1-9c61-481ca568d022","Type":"ContainerDied","Data":"13189da41e0fb30fa7cca9718222038a2b578d40c4f21c5e350b74e753b85587"} Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.042560 4778 scope.go:117] "RemoveContainer" containerID="718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.042741 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xksl"] Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.045098 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xksl"] Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.058593 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f438f2a3-60c0-4554-a49b-030545f8139c" (UID: "f438f2a3-60c0-4554-a49b-030545f8139c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.086019 4778 scope.go:117] "RemoveContainer" containerID="167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.107411 4778 scope.go:117] "RemoveContainer" containerID="1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592" Mar 12 13:17:02 crc kubenswrapper[4778]: E0312 13:17:02.107957 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592\": container with ID starting with 1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592 not found: ID does not exist" containerID="1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.107988 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592"} err="failed to get container status \"1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592\": rpc error: code = NotFound desc = could not find container \"1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592\": container with ID starting with 1dc3137ddc227e6024fccc0afbe6f1d93623b9e53c63a937c6719203e66ee592 not found: ID does not exist" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.108012 4778 scope.go:117] "RemoveContainer" containerID="718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba" Mar 12 13:17:02 crc kubenswrapper[4778]: E0312 13:17:02.108346 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba\": container with ID starting with 718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba not found: ID does not exist" containerID="718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.108371 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba"} err="failed to get container status \"718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba\": rpc error: code = NotFound desc = could not find container \"718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba\": container with ID starting with 718ec5d1f6755df76f8300b916ef0eb0663019d9610ddce44e4b950ef7dec3ba not found: ID does not exist" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.108384 4778 scope.go:117] "RemoveContainer" containerID="167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252" Mar 12 13:17:02 crc kubenswrapper[4778]: E0312 13:17:02.108634 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252\": container with ID starting with 167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252 not found: ID does not exist" containerID="167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.108653 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252"} err="failed to get container status \"167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252\": rpc error: code = NotFound desc = could not find container \"167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252\": container with ID starting with 167b98bcb75be92dcb64515712bdd5c31feb59c13d9a61d37d29e56c03f4a252 not found: ID does not exist" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.108667 4778 scope.go:117] "RemoveContainer" containerID="06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.119868 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f438f2a3-60c0-4554-a49b-030545f8139c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.122884 4778 scope.go:117] "RemoveContainer" containerID="ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.162702 4778 scope.go:117] "RemoveContainer" containerID="9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.186592 4778 scope.go:117] "RemoveContainer" containerID="06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd" Mar 12 13:17:02 crc kubenswrapper[4778]: E0312 13:17:02.186979 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd\": container with ID starting with 06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd not found: ID does not exist" containerID="06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.187017 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd"} err="failed to get container status \"06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd\": rpc error: code = NotFound desc = could not find container \"06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd\": container with ID starting with 06e3e529ea6d479f93a4c0f8dc62611d7db0ca000158fd5d08aba4b4784ec2bd not found: ID does not exist" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.187042 4778 scope.go:117] "RemoveContainer" containerID="ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354" Mar 12 13:17:02 crc kubenswrapper[4778]: E0312 13:17:02.187423 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354\": container with ID starting with ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354 not found: ID does not exist" containerID="ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.187438 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354"} err="failed to get container status \"ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354\": rpc error: code = NotFound desc = could not find container \"ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354\": container with ID starting with ac4be4ba9c0f65056e92751bc6e83a1871b4710d28a0b4f32b544fe6c70e1354 not found: ID does not exist" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.187450 4778 scope.go:117] "RemoveContainer" containerID="9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e" Mar 12 13:17:02 crc kubenswrapper[4778]: E0312 13:17:02.187776 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e\": container with ID starting with 9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e not found: ID does not exist" containerID="9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.187791 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e"} err="failed to get container status \"9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e\": rpc error: code = NotFound desc = could not find container \"9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e\": container with ID starting with 9727ee6f8e8c78a7a69962ee912839b2519b88f461321e8f43bb35e450713d1e not found: ID does not exist" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.260049 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" path="/var/lib/kubelet/pods/de4557b4-7957-47a0-8c42-845be1fa0f32/volumes" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.367209 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.382602 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5s5vs"] Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.382896 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.386595 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5s5vs"] Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.391520 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524100 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-utilities\") pod \"1d185732-cd6b-44c6-b4db-ee9ade00c683\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524550 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-operator-metrics\") pod \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524584 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6mx5\" (UniqueName: \"kubernetes.io/projected/651601bd-18fe-4ca1-9c61-481ca568d022-kube-api-access-n6mx5\") pod \"651601bd-18fe-4ca1-9c61-481ca568d022\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524617 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzdp9\" (UniqueName: \"kubernetes.io/projected/1d185732-cd6b-44c6-b4db-ee9ade00c683-kube-api-access-zzdp9\") pod \"1d185732-cd6b-44c6-b4db-ee9ade00c683\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524645 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-trusted-ca\") pod \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524663 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-catalog-content\") pod \"651601bd-18fe-4ca1-9c61-481ca568d022\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524702 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-utilities\") pod \"651601bd-18fe-4ca1-9c61-481ca568d022\" (UID: \"651601bd-18fe-4ca1-9c61-481ca568d022\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524733 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmb4r\" (UniqueName: \"kubernetes.io/projected/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-kube-api-access-pmb4r\") pod \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\" (UID: \"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.524760 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-catalog-content\") pod \"1d185732-cd6b-44c6-b4db-ee9ade00c683\" (UID: \"1d185732-cd6b-44c6-b4db-ee9ade00c683\") " Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.525214 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-utilities" (OuterVolumeSpecName: "utilities") pod "1d185732-cd6b-44c6-b4db-ee9ade00c683" (UID: "1d185732-cd6b-44c6-b4db-ee9ade00c683"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.526127 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" (UID: "24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.526688 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-utilities" (OuterVolumeSpecName: "utilities") pod "651601bd-18fe-4ca1-9c61-481ca568d022" (UID: "651601bd-18fe-4ca1-9c61-481ca568d022"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.527326 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/651601bd-18fe-4ca1-9c61-481ca568d022-kube-api-access-n6mx5" (OuterVolumeSpecName: "kube-api-access-n6mx5") pod "651601bd-18fe-4ca1-9c61-481ca568d022" (UID: "651601bd-18fe-4ca1-9c61-481ca568d022"). InnerVolumeSpecName "kube-api-access-n6mx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.528616 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-kube-api-access-pmb4r" (OuterVolumeSpecName: "kube-api-access-pmb4r") pod "24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" (UID: "24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d"). InnerVolumeSpecName "kube-api-access-pmb4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.528661 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d185732-cd6b-44c6-b4db-ee9ade00c683-kube-api-access-zzdp9" (OuterVolumeSpecName: "kube-api-access-zzdp9") pod "1d185732-cd6b-44c6-b4db-ee9ade00c683" (UID: "1d185732-cd6b-44c6-b4db-ee9ade00c683"). InnerVolumeSpecName "kube-api-access-zzdp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.529444 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" (UID: "24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.587232 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "651601bd-18fe-4ca1-9c61-481ca568d022" (UID: "651601bd-18fe-4ca1-9c61-481ca568d022"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.599439 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d185732-cd6b-44c6-b4db-ee9ade00c683" (UID: "1d185732-cd6b-44c6-b4db-ee9ade00c683"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626142 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmb4r\" (UniqueName: \"kubernetes.io/projected/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-kube-api-access-pmb4r\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626176 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626197 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d185732-cd6b-44c6-b4db-ee9ade00c683-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626206 4778 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626216 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6mx5\" (UniqueName: \"kubernetes.io/projected/651601bd-18fe-4ca1-9c61-481ca568d022-kube-api-access-n6mx5\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626229 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzdp9\" (UniqueName: \"kubernetes.io/projected/1d185732-cd6b-44c6-b4db-ee9ade00c683-kube-api-access-zzdp9\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626241 4778 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626252 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:02 crc kubenswrapper[4778]: I0312 13:17:02.626259 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/651601bd-18fe-4ca1-9c61-481ca568d022-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.031901 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qx9d8" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.031919 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qx9d8" event={"ID":"651601bd-18fe-4ca1-9c61-481ca568d022","Type":"ContainerDied","Data":"592ed663fa0a363547ba9675a7740b1982ac31820675fa1bc6b541164ee13dff"} Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.032072 4778 scope.go:117] "RemoveContainer" containerID="13189da41e0fb30fa7cca9718222038a2b578d40c4f21c5e350b74e753b85587" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.038819 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khr6h" event={"ID":"1d185732-cd6b-44c6-b4db-ee9ade00c683","Type":"ContainerDied","Data":"f4257f2b5ae0b8d1695cb20eed3d7af4ca3c14b5f906e52fd4e46f8237158ff5"} Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.039039 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khr6h" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.040765 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" event={"ID":"3b062c23-5acd-430d-aa6c-24b48a725594","Type":"ContainerStarted","Data":"d197f824483b8e30c9baf11e1ffc2173dbd46d1ed31a6528325bfce5da893206"} Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.040984 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.043438 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" event={"ID":"24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d","Type":"ContainerDied","Data":"9ea9ce91a5458d09f7e543bf678a01cfeb2e8462d6860a8c5523bea49359f807"} Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.043567 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2wqm5" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.044259 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.067614 4778 scope.go:117] "RemoveContainer" containerID="777dcb7d13b3c9f17ff760e883a8a2c8d277b3c6622f9924b38301e80f9b85e9" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.070057 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hvmk8" podStartSLOduration=2.070018221 podStartE2EDuration="2.070018221s" podCreationTimestamp="2026-03-12 13:17:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:17:03.061955474 +0000 UTC m=+441.510650900" watchObservedRunningTime="2026-03-12 13:17:03.070018221 +0000 UTC m=+441.518713617" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.102996 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2wqm5"] Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.107491 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2wqm5"] Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.116785 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qx9d8"] Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.130634 4778 scope.go:117] "RemoveContainer" containerID="768c08538cc35f7dca92094b0ee56f8d00acc523e23bc32165393cb6d17f7cd2" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.132650 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qx9d8"] Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.135849 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-khr6h"] Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.139090 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-khr6h"] Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.145683 4778 scope.go:117] "RemoveContainer" containerID="b352e6584b478e7228a408cc5d6c8b18473e75a0de7be819c32ae9b98a707a4e" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.160817 4778 scope.go:117] "RemoveContainer" containerID="84fe3c954d7e0d1d6303467d2621bf3b31d896882603252deb19491a2fa354ed" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.176835 4778 scope.go:117] "RemoveContainer" containerID="05d961ad3b7bd74a33e24a693f2775dd8f5c4483b25df2fe323f0e88cb5ff934" Mar 12 13:17:03 crc kubenswrapper[4778]: I0312 13:17:03.192303 4778 scope.go:117] "RemoveContainer" containerID="013c13acbd136a9ae3c6c39b9470a59aa4ab705637939d6af761af9e92e81b9c" Mar 12 13:17:04 crc kubenswrapper[4778]: I0312 13:17:04.262073 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" path="/var/lib/kubelet/pods/1d185732-cd6b-44c6-b4db-ee9ade00c683/volumes" Mar 12 13:17:04 crc kubenswrapper[4778]: I0312 13:17:04.263117 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" path="/var/lib/kubelet/pods/24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d/volumes" Mar 12 13:17:04 crc kubenswrapper[4778]: I0312 13:17:04.263906 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" path="/var/lib/kubelet/pods/651601bd-18fe-4ca1-9c61-481ca568d022/volumes" Mar 12 13:17:04 crc kubenswrapper[4778]: I0312 13:17:04.265600 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" path="/var/lib/kubelet/pods/f438f2a3-60c0-4554-a49b-030545f8139c/volumes" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331211 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k57lm"] Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331503 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331522 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331540 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331552 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331568 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331579 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331595 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331606 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331619 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331630 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331643 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331653 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331670 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331680 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331694 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" containerName="marketplace-operator" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331704 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" containerName="marketplace-operator" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331721 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331732 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331746 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331757 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331772 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331783 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331803 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331814 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="extract-content" Mar 12 13:17:05 crc kubenswrapper[4778]: E0312 13:17:05.331833 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331844 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="extract-utilities" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.331988 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="651601bd-18fe-4ca1-9c61-481ca568d022" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.332008 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f4aaf5-c17b-4cd8-9284-6df37f1c2f2d" containerName="marketplace-operator" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.332027 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f438f2a3-60c0-4554-a49b-030545f8139c" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.332040 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d185732-cd6b-44c6-b4db-ee9ade00c683" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.332063 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4557b4-7957-47a0-8c42-845be1fa0f32" containerName="registry-server" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.335298 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.339984 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.349095 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k57lm"] Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.370473 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-utilities\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.370547 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqdwm\" (UniqueName: \"kubernetes.io/projected/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-kube-api-access-jqdwm\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.370661 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-catalog-content\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.471113 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-utilities\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.471227 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqdwm\" (UniqueName: \"kubernetes.io/projected/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-kube-api-access-jqdwm\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.471340 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-catalog-content\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.471843 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-utilities\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.472205 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-catalog-content\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.495517 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqdwm\" (UniqueName: \"kubernetes.io/projected/1d67fa18-822d-4685-a7a1-5b8b8c39c96a-kube-api-access-jqdwm\") pod \"redhat-marketplace-k57lm\" (UID: \"1d67fa18-822d-4685-a7a1-5b8b8c39c96a\") " pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.528375 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r99nz"] Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.529313 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.531564 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.548407 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r99nz"] Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.667595 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.672692 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mccx\" (UniqueName: \"kubernetes.io/projected/89b39891-5207-4289-807f-57d00acb2937-kube-api-access-7mccx\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.672762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-utilities\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.672809 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-catalog-content\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.774087 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mccx\" (UniqueName: \"kubernetes.io/projected/89b39891-5207-4289-807f-57d00acb2937-kube-api-access-7mccx\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.774214 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-utilities\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.774264 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-catalog-content\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.774853 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-catalog-content\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.775158 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-utilities\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.800766 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mccx\" (UniqueName: \"kubernetes.io/projected/89b39891-5207-4289-807f-57d00acb2937-kube-api-access-7mccx\") pod \"redhat-operators-r99nz\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:05 crc kubenswrapper[4778]: I0312 13:17:05.847742 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:06 crc kubenswrapper[4778]: I0312 13:17:06.090467 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k57lm"] Mar 12 13:17:06 crc kubenswrapper[4778]: W0312 13:17:06.097299 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d67fa18_822d_4685_a7a1_5b8b8c39c96a.slice/crio-e915930a5cfb75a4b06093def3f48df6ba60a7c8483b7b284c420a1c0656b5df WatchSource:0}: Error finding container e915930a5cfb75a4b06093def3f48df6ba60a7c8483b7b284c420a1c0656b5df: Status 404 returned error can't find the container with id e915930a5cfb75a4b06093def3f48df6ba60a7c8483b7b284c420a1c0656b5df Mar 12 13:17:06 crc kubenswrapper[4778]: I0312 13:17:06.213931 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r99nz"] Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.083838 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k57lm" event={"ID":"1d67fa18-822d-4685-a7a1-5b8b8c39c96a","Type":"ContainerDied","Data":"b163b5c1861d00221f3bbf389e3409b5816c14f6a5e512a09b8553e7bfb2f484"} Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.083646 4778 generic.go:334] "Generic (PLEG): container finished" podID="1d67fa18-822d-4685-a7a1-5b8b8c39c96a" containerID="b163b5c1861d00221f3bbf389e3409b5816c14f6a5e512a09b8553e7bfb2f484" exitCode=0 Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.085293 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k57lm" event={"ID":"1d67fa18-822d-4685-a7a1-5b8b8c39c96a","Type":"ContainerStarted","Data":"e915930a5cfb75a4b06093def3f48df6ba60a7c8483b7b284c420a1c0656b5df"} Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.088910 4778 generic.go:334] "Generic (PLEG): container finished" podID="89b39891-5207-4289-807f-57d00acb2937" containerID="984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b" exitCode=0 Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.088943 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99nz" event={"ID":"89b39891-5207-4289-807f-57d00acb2937","Type":"ContainerDied","Data":"984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b"} Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.088964 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99nz" event={"ID":"89b39891-5207-4289-807f-57d00acb2937","Type":"ContainerStarted","Data":"d44d2bee1e2b4ddf99f45277d9dc014b3b21712ebf48e47cb48538e60ac5ff80"} Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.729663 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fhcbf"] Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.730805 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.733774 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.743964 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fhcbf"] Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.898897 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-catalog-content\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.898951 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2v5x\" (UniqueName: \"kubernetes.io/projected/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-kube-api-access-s2v5x\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.899519 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-utilities\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.933789 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-scbxn"] Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.935089 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.937681 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 13:17:07 crc kubenswrapper[4778]: I0312 13:17:07.943053 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-scbxn"] Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.000695 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-catalog-content\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.000746 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2v5x\" (UniqueName: \"kubernetes.io/projected/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-kube-api-access-s2v5x\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.000936 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-utilities\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.001163 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-catalog-content\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.001551 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-utilities\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.020504 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2v5x\" (UniqueName: \"kubernetes.io/projected/b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef-kube-api-access-s2v5x\") pod \"certified-operators-fhcbf\" (UID: \"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef\") " pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.063413 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.101831 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-catalog-content\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.102133 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpxbz\" (UniqueName: \"kubernetes.io/projected/f2f91915-3841-4662-88e4-82a22df0b131-kube-api-access-xpxbz\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.102158 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-utilities\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.202943 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-catalog-content\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.203298 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpxbz\" (UniqueName: \"kubernetes.io/projected/f2f91915-3841-4662-88e4-82a22df0b131-kube-api-access-xpxbz\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.203325 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-utilities\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.203548 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-catalog-content\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.203697 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-utilities\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.239585 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpxbz\" (UniqueName: \"kubernetes.io/projected/f2f91915-3841-4662-88e4-82a22df0b131-kube-api-access-xpxbz\") pod \"community-operators-scbxn\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.248921 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.500891 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fhcbf"] Mar 12 13:17:08 crc kubenswrapper[4778]: W0312 13:17:08.510998 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5b1dff9_c32b_4a91_863c_10b5ea4bc4ef.slice/crio-b17915a7a54db4e4374c5e1c27b5ab3e9dfdfb65aa22034620c6839ba47335e8 WatchSource:0}: Error finding container b17915a7a54db4e4374c5e1c27b5ab3e9dfdfb65aa22034620c6839ba47335e8: Status 404 returned error can't find the container with id b17915a7a54db4e4374c5e1c27b5ab3e9dfdfb65aa22034620c6839ba47335e8 Mar 12 13:17:08 crc kubenswrapper[4778]: I0312 13:17:08.637581 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-scbxn"] Mar 12 13:17:08 crc kubenswrapper[4778]: W0312 13:17:08.687898 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2f91915_3841_4662_88e4_82a22df0b131.slice/crio-a19d957b7ce97a07ce6c0132cc4944c7bac635fad459f95aeb77803a9db2f905 WatchSource:0}: Error finding container a19d957b7ce97a07ce6c0132cc4944c7bac635fad459f95aeb77803a9db2f905: Status 404 returned error can't find the container with id a19d957b7ce97a07ce6c0132cc4944c7bac635fad459f95aeb77803a9db2f905 Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.109325 4778 generic.go:334] "Generic (PLEG): container finished" podID="89b39891-5207-4289-807f-57d00acb2937" containerID="805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e" exitCode=0 Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.109893 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99nz" event={"ID":"89b39891-5207-4289-807f-57d00acb2937","Type":"ContainerDied","Data":"805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e"} Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.115172 4778 generic.go:334] "Generic (PLEG): container finished" podID="b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef" containerID="cbc2b1582a7a0850032095cce0edc23160709ce6369efa26cb488729d09bf744" exitCode=0 Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.115759 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhcbf" event={"ID":"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef","Type":"ContainerDied","Data":"cbc2b1582a7a0850032095cce0edc23160709ce6369efa26cb488729d09bf744"} Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.115794 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhcbf" event={"ID":"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef","Type":"ContainerStarted","Data":"b17915a7a54db4e4374c5e1c27b5ab3e9dfdfb65aa22034620c6839ba47335e8"} Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.117724 4778 generic.go:334] "Generic (PLEG): container finished" podID="f2f91915-3841-4662-88e4-82a22df0b131" containerID="10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62" exitCode=0 Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.117751 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerDied","Data":"10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62"} Mar 12 13:17:09 crc kubenswrapper[4778]: I0312 13:17:09.117776 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerStarted","Data":"a19d957b7ce97a07ce6c0132cc4944c7bac635fad459f95aeb77803a9db2f905"} Mar 12 13:17:10 crc kubenswrapper[4778]: I0312 13:17:10.136839 4778 generic.go:334] "Generic (PLEG): container finished" podID="1d67fa18-822d-4685-a7a1-5b8b8c39c96a" containerID="a12f9320b89f1cff13da61cda5d0d3172add5ff7c7b42ca5c292f58ca0c519d2" exitCode=0 Mar 12 13:17:10 crc kubenswrapper[4778]: I0312 13:17:10.136995 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k57lm" event={"ID":"1d67fa18-822d-4685-a7a1-5b8b8c39c96a","Type":"ContainerDied","Data":"a12f9320b89f1cff13da61cda5d0d3172add5ff7c7b42ca5c292f58ca0c519d2"} Mar 12 13:17:10 crc kubenswrapper[4778]: I0312 13:17:10.141743 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerStarted","Data":"814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e"} Mar 12 13:17:10 crc kubenswrapper[4778]: I0312 13:17:10.147173 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99nz" event={"ID":"89b39891-5207-4289-807f-57d00acb2937","Type":"ContainerStarted","Data":"712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31"} Mar 12 13:17:10 crc kubenswrapper[4778]: I0312 13:17:10.196711 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r99nz" podStartSLOduration=2.715893827 podStartE2EDuration="5.196693017s" podCreationTimestamp="2026-03-12 13:17:05 +0000 UTC" firstStartedPulling="2026-03-12 13:17:07.090877234 +0000 UTC m=+445.539572630" lastFinishedPulling="2026-03-12 13:17:09.571676414 +0000 UTC m=+448.020371820" observedRunningTime="2026-03-12 13:17:10.178524391 +0000 UTC m=+448.627219787" watchObservedRunningTime="2026-03-12 13:17:10.196693017 +0000 UTC m=+448.645388433" Mar 12 13:17:11 crc kubenswrapper[4778]: I0312 13:17:11.154778 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k57lm" event={"ID":"1d67fa18-822d-4685-a7a1-5b8b8c39c96a","Type":"ContainerStarted","Data":"344d2b901b2303d0a19f1c46054188f6685e0524047e61f31c15e844234ba822"} Mar 12 13:17:11 crc kubenswrapper[4778]: I0312 13:17:11.157709 4778 generic.go:334] "Generic (PLEG): container finished" podID="b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef" containerID="4e89930cc55f42c1b9c607ed232d72bcdba978dafb0fafb2584a55b43f8b32a4" exitCode=0 Mar 12 13:17:11 crc kubenswrapper[4778]: I0312 13:17:11.157803 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhcbf" event={"ID":"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef","Type":"ContainerDied","Data":"4e89930cc55f42c1b9c607ed232d72bcdba978dafb0fafb2584a55b43f8b32a4"} Mar 12 13:17:11 crc kubenswrapper[4778]: I0312 13:17:11.160980 4778 generic.go:334] "Generic (PLEG): container finished" podID="f2f91915-3841-4662-88e4-82a22df0b131" containerID="814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e" exitCode=0 Mar 12 13:17:11 crc kubenswrapper[4778]: I0312 13:17:11.161538 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerDied","Data":"814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e"} Mar 12 13:17:11 crc kubenswrapper[4778]: I0312 13:17:11.182397 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k57lm" podStartSLOduration=2.688240841 podStartE2EDuration="6.182379198s" podCreationTimestamp="2026-03-12 13:17:05 +0000 UTC" firstStartedPulling="2026-03-12 13:17:07.087509701 +0000 UTC m=+445.536205117" lastFinishedPulling="2026-03-12 13:17:10.581648078 +0000 UTC m=+449.030343474" observedRunningTime="2026-03-12 13:17:11.181479741 +0000 UTC m=+449.630175137" watchObservedRunningTime="2026-03-12 13:17:11.182379198 +0000 UTC m=+449.631074594" Mar 12 13:17:12 crc kubenswrapper[4778]: I0312 13:17:12.179710 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhcbf" event={"ID":"b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef","Type":"ContainerStarted","Data":"30cfba142743b15f9002d2a0931af09174d76a9d4edab7d4eecf93d5cb1c7403"} Mar 12 13:17:12 crc kubenswrapper[4778]: I0312 13:17:12.194453 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerStarted","Data":"b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76"} Mar 12 13:17:12 crc kubenswrapper[4778]: I0312 13:17:12.221661 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fhcbf" podStartSLOduration=2.772273599 podStartE2EDuration="5.221641498s" podCreationTimestamp="2026-03-12 13:17:07 +0000 UTC" firstStartedPulling="2026-03-12 13:17:09.116589479 +0000 UTC m=+447.565284875" lastFinishedPulling="2026-03-12 13:17:11.565957378 +0000 UTC m=+450.014652774" observedRunningTime="2026-03-12 13:17:12.202162843 +0000 UTC m=+450.650858259" watchObservedRunningTime="2026-03-12 13:17:12.221641498 +0000 UTC m=+450.670336904" Mar 12 13:17:12 crc kubenswrapper[4778]: I0312 13:17:12.222620 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-scbxn" podStartSLOduration=2.779431728 podStartE2EDuration="5.222610168s" podCreationTimestamp="2026-03-12 13:17:07 +0000 UTC" firstStartedPulling="2026-03-12 13:17:09.119699824 +0000 UTC m=+447.568395220" lastFinishedPulling="2026-03-12 13:17:11.562878264 +0000 UTC m=+450.011573660" observedRunningTime="2026-03-12 13:17:12.219551814 +0000 UTC m=+450.668247210" watchObservedRunningTime="2026-03-12 13:17:12.222610168 +0000 UTC m=+450.671305584" Mar 12 13:17:15 crc kubenswrapper[4778]: I0312 13:17:15.668835 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:15 crc kubenswrapper[4778]: I0312 13:17:15.670739 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:15 crc kubenswrapper[4778]: I0312 13:17:15.715003 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:15 crc kubenswrapper[4778]: I0312 13:17:15.849062 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:15 crc kubenswrapper[4778]: I0312 13:17:15.849383 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:16 crc kubenswrapper[4778]: I0312 13:17:16.278977 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k57lm" Mar 12 13:17:16 crc kubenswrapper[4778]: I0312 13:17:16.895441 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r99nz" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="registry-server" probeResult="failure" output=< Mar 12 13:17:16 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:17:16 crc kubenswrapper[4778]: > Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.064491 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.064545 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.119259 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.250252 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.251464 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.281474 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fhcbf" Mar 12 13:17:18 crc kubenswrapper[4778]: I0312 13:17:18.313139 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:19 crc kubenswrapper[4778]: I0312 13:17:19.295977 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.177704 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" podUID="51ee714f-fb23-4420-9e70-1b3134eea18e" containerName="registry" containerID="cri-o://29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111" gracePeriod=30 Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.586179 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653090 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51ee714f-fb23-4420-9e70-1b3134eea18e-ca-trust-extracted\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653209 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-bound-sa-token\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653242 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtgbd\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-kube-api-access-mtgbd\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653301 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51ee714f-fb23-4420-9e70-1b3134eea18e-installation-pull-secrets\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653336 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-trusted-ca\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653378 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-tls\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653647 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.653698 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-certificates\") pod \"51ee714f-fb23-4420-9e70-1b3134eea18e\" (UID: \"51ee714f-fb23-4420-9e70-1b3134eea18e\") " Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.654219 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.654464 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.660279 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51ee714f-fb23-4420-9e70-1b3134eea18e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.662364 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-kube-api-access-mtgbd" (OuterVolumeSpecName: "kube-api-access-mtgbd") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "kube-api-access-mtgbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.663752 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.664214 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.667696 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.668774 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ee714f-fb23-4420-9e70-1b3134eea18e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "51ee714f-fb23-4420-9e70-1b3134eea18e" (UID: "51ee714f-fb23-4420-9e70-1b3134eea18e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755725 4778 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755780 4778 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51ee714f-fb23-4420-9e70-1b3134eea18e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755796 4778 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755805 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtgbd\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-kube-api-access-mtgbd\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755818 4778 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51ee714f-fb23-4420-9e70-1b3134eea18e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755829 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51ee714f-fb23-4420-9e70-1b3134eea18e-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:24 crc kubenswrapper[4778]: I0312 13:17:24.755839 4778 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51ee714f-fb23-4420-9e70-1b3134eea18e-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.272880 4778 generic.go:334] "Generic (PLEG): container finished" podID="51ee714f-fb23-4420-9e70-1b3134eea18e" containerID="29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111" exitCode=0 Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.272916 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" event={"ID":"51ee714f-fb23-4420-9e70-1b3134eea18e","Type":"ContainerDied","Data":"29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111"} Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.272947 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" event={"ID":"51ee714f-fb23-4420-9e70-1b3134eea18e","Type":"ContainerDied","Data":"9fdf9bc3368d582f75afb64ef1bd7b59c9e3cd5fe63a9b2265425474dba3a3b4"} Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.272967 4778 scope.go:117] "RemoveContainer" containerID="29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.272969 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxrx4" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.302806 4778 scope.go:117] "RemoveContainer" containerID="29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.305137 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxrx4"] Mar 12 13:17:25 crc kubenswrapper[4778]: E0312 13:17:25.306087 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111\": container with ID starting with 29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111 not found: ID does not exist" containerID="29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.306144 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111"} err="failed to get container status \"29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111\": rpc error: code = NotFound desc = could not find container \"29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111\": container with ID starting with 29df7c95c025412716ae854e04324e43fa3cc12e2e3e9061ce1a3a4518451111 not found: ID does not exist" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.311238 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxrx4"] Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.888062 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:25 crc kubenswrapper[4778]: I0312 13:17:25.928359 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 13:17:26 crc kubenswrapper[4778]: I0312 13:17:26.260238 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ee714f-fb23-4420-9e70-1b3134eea18e" path="/var/lib/kubelet/pods/51ee714f-fb23-4420-9e70-1b3134eea18e/volumes" Mar 12 13:17:32 crc kubenswrapper[4778]: I0312 13:17:32.355871 4778 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podf438f2a3-60c0-4554-a49b-030545f8139c"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podf438f2a3-60c0-4554-a49b-030545f8139c] : Timed out while waiting for systemd to remove kubepods-burstable-podf438f2a3_60c0_4554_a49b_030545f8139c.slice" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.132330 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555358-txmvp"] Mar 12 13:18:00 crc kubenswrapper[4778]: E0312 13:18:00.133044 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ee714f-fb23-4420-9e70-1b3134eea18e" containerName="registry" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.133059 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ee714f-fb23-4420-9e70-1b3134eea18e" containerName="registry" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.133206 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ee714f-fb23-4420-9e70-1b3134eea18e" containerName="registry" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.133649 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.137619 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.137855 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.137866 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.142879 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555358-txmvp"] Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.282220 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcrf5\" (UniqueName: \"kubernetes.io/projected/61c6485d-2d53-47d9-866a-31eb90ac254e-kube-api-access-mcrf5\") pod \"auto-csr-approver-29555358-txmvp\" (UID: \"61c6485d-2d53-47d9-866a-31eb90ac254e\") " pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.384094 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcrf5\" (UniqueName: \"kubernetes.io/projected/61c6485d-2d53-47d9-866a-31eb90ac254e-kube-api-access-mcrf5\") pod \"auto-csr-approver-29555358-txmvp\" (UID: \"61c6485d-2d53-47d9-866a-31eb90ac254e\") " pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.405159 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcrf5\" (UniqueName: \"kubernetes.io/projected/61c6485d-2d53-47d9-866a-31eb90ac254e-kube-api-access-mcrf5\") pod \"auto-csr-approver-29555358-txmvp\" (UID: \"61c6485d-2d53-47d9-866a-31eb90ac254e\") " pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.452826 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.841493 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555358-txmvp"] Mar 12 13:18:00 crc kubenswrapper[4778]: I0312 13:18:00.852771 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:18:01 crc kubenswrapper[4778]: I0312 13:18:01.490046 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555358-txmvp" event={"ID":"61c6485d-2d53-47d9-866a-31eb90ac254e","Type":"ContainerStarted","Data":"4310735102e019aa75c378b665ecbac2727e33dd70292149fff77566559719ef"} Mar 12 13:18:03 crc kubenswrapper[4778]: I0312 13:18:03.507840 4778 generic.go:334] "Generic (PLEG): container finished" podID="61c6485d-2d53-47d9-866a-31eb90ac254e" containerID="42a7fef965fea72fd4ae8fcc7e99e6b821d3626af8cb88a527c7193c956003a6" exitCode=0 Mar 12 13:18:03 crc kubenswrapper[4778]: I0312 13:18:03.507904 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555358-txmvp" event={"ID":"61c6485d-2d53-47d9-866a-31eb90ac254e","Type":"ContainerDied","Data":"42a7fef965fea72fd4ae8fcc7e99e6b821d3626af8cb88a527c7193c956003a6"} Mar 12 13:18:04 crc kubenswrapper[4778]: I0312 13:18:04.828727 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:04 crc kubenswrapper[4778]: I0312 13:18:04.936596 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcrf5\" (UniqueName: \"kubernetes.io/projected/61c6485d-2d53-47d9-866a-31eb90ac254e-kube-api-access-mcrf5\") pod \"61c6485d-2d53-47d9-866a-31eb90ac254e\" (UID: \"61c6485d-2d53-47d9-866a-31eb90ac254e\") " Mar 12 13:18:04 crc kubenswrapper[4778]: I0312 13:18:04.941530 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c6485d-2d53-47d9-866a-31eb90ac254e-kube-api-access-mcrf5" (OuterVolumeSpecName: "kube-api-access-mcrf5") pod "61c6485d-2d53-47d9-866a-31eb90ac254e" (UID: "61c6485d-2d53-47d9-866a-31eb90ac254e"). InnerVolumeSpecName "kube-api-access-mcrf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:18:05 crc kubenswrapper[4778]: I0312 13:18:05.037854 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcrf5\" (UniqueName: \"kubernetes.io/projected/61c6485d-2d53-47d9-866a-31eb90ac254e-kube-api-access-mcrf5\") on node \"crc\" DevicePath \"\"" Mar 12 13:18:05 crc kubenswrapper[4778]: I0312 13:18:05.524945 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555358-txmvp" event={"ID":"61c6485d-2d53-47d9-866a-31eb90ac254e","Type":"ContainerDied","Data":"4310735102e019aa75c378b665ecbac2727e33dd70292149fff77566559719ef"} Mar 12 13:18:05 crc kubenswrapper[4778]: I0312 13:18:05.524997 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4310735102e019aa75c378b665ecbac2727e33dd70292149fff77566559719ef" Mar 12 13:18:05 crc kubenswrapper[4778]: I0312 13:18:05.525013 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555358-txmvp" Mar 12 13:18:05 crc kubenswrapper[4778]: I0312 13:18:05.904377 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555352-q7fvr"] Mar 12 13:18:05 crc kubenswrapper[4778]: I0312 13:18:05.912418 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555352-q7fvr"] Mar 12 13:18:06 crc kubenswrapper[4778]: I0312 13:18:06.259582 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f210efd-2ac0-4b67-89c5-fcd9f52f6e01" path="/var/lib/kubelet/pods/9f210efd-2ac0-4b67-89c5-fcd9f52f6e01/volumes" Mar 12 13:18:58 crc kubenswrapper[4778]: I0312 13:18:58.557387 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:18:58 crc kubenswrapper[4778]: I0312 13:18:58.557984 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:19:28 crc kubenswrapper[4778]: I0312 13:19:28.557675 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:19:28 crc kubenswrapper[4778]: I0312 13:19:28.558263 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:19:46 crc kubenswrapper[4778]: I0312 13:19:46.568178 4778 scope.go:117] "RemoveContainer" containerID="5ab6ab1e87e3d9a4f7941a7ab56868950f541c7821fdd08fb7b7e95206f0cb25" Mar 12 13:19:58 crc kubenswrapper[4778]: I0312 13:19:58.558075 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:19:58 crc kubenswrapper[4778]: I0312 13:19:58.558521 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:19:58 crc kubenswrapper[4778]: I0312 13:19:58.558597 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:19:58 crc kubenswrapper[4778]: I0312 13:19:58.559449 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e50690e6aff1fd408e6201d1eee1240e692ce04bc21873dbbe85a5f2d638d704"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:19:58 crc kubenswrapper[4778]: I0312 13:19:58.559542 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://e50690e6aff1fd408e6201d1eee1240e692ce04bc21873dbbe85a5f2d638d704" gracePeriod=600 Mar 12 13:19:59 crc kubenswrapper[4778]: I0312 13:19:59.212926 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="e50690e6aff1fd408e6201d1eee1240e692ce04bc21873dbbe85a5f2d638d704" exitCode=0 Mar 12 13:19:59 crc kubenswrapper[4778]: I0312 13:19:59.213179 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"e50690e6aff1fd408e6201d1eee1240e692ce04bc21873dbbe85a5f2d638d704"} Mar 12 13:19:59 crc kubenswrapper[4778]: I0312 13:19:59.213256 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"dfcc37339849724c4aacca3262255dd43897a2284c2172380a90cc97f52e3a46"} Mar 12 13:19:59 crc kubenswrapper[4778]: I0312 13:19:59.213317 4778 scope.go:117] "RemoveContainer" containerID="dcabd48eda797c052967d086d455193bf30a1f05151385a52352d733c58148f7" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.142770 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555360-vwflx"] Mar 12 13:20:00 crc kubenswrapper[4778]: E0312 13:20:00.143031 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c6485d-2d53-47d9-866a-31eb90ac254e" containerName="oc" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.143046 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c6485d-2d53-47d9-866a-31eb90ac254e" containerName="oc" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.143173 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c6485d-2d53-47d9-866a-31eb90ac254e" containerName="oc" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.143648 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.146701 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.147172 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.147465 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.153584 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555360-vwflx"] Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.228877 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzwxv\" (UniqueName: \"kubernetes.io/projected/4c617404-7840-495c-80da-593af33f77d6-kube-api-access-kzwxv\") pod \"auto-csr-approver-29555360-vwflx\" (UID: \"4c617404-7840-495c-80da-593af33f77d6\") " pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.330203 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzwxv\" (UniqueName: \"kubernetes.io/projected/4c617404-7840-495c-80da-593af33f77d6-kube-api-access-kzwxv\") pod \"auto-csr-approver-29555360-vwflx\" (UID: \"4c617404-7840-495c-80da-593af33f77d6\") " pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.354070 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzwxv\" (UniqueName: \"kubernetes.io/projected/4c617404-7840-495c-80da-593af33f77d6-kube-api-access-kzwxv\") pod \"auto-csr-approver-29555360-vwflx\" (UID: \"4c617404-7840-495c-80da-593af33f77d6\") " pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.512695 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:00 crc kubenswrapper[4778]: I0312 13:20:00.711454 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555360-vwflx"] Mar 12 13:20:01 crc kubenswrapper[4778]: I0312 13:20:01.229064 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555360-vwflx" event={"ID":"4c617404-7840-495c-80da-593af33f77d6","Type":"ContainerStarted","Data":"0351e8b2cc8d22e69dee8334ff2e34916d4f30011467286d86f3b0031f613676"} Mar 12 13:20:02 crc kubenswrapper[4778]: I0312 13:20:02.234920 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555360-vwflx" event={"ID":"4c617404-7840-495c-80da-593af33f77d6","Type":"ContainerStarted","Data":"97b3a747ac158c0518500113b5af025bff04e06faaee081df03d1a06860f190f"} Mar 12 13:20:02 crc kubenswrapper[4778]: I0312 13:20:02.250110 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555360-vwflx" podStartSLOduration=1.049887438 podStartE2EDuration="2.250087831s" podCreationTimestamp="2026-03-12 13:20:00 +0000 UTC" firstStartedPulling="2026-03-12 13:20:00.716846248 +0000 UTC m=+619.165541644" lastFinishedPulling="2026-03-12 13:20:01.917046601 +0000 UTC m=+620.365742037" observedRunningTime="2026-03-12 13:20:02.247946508 +0000 UTC m=+620.696641964" watchObservedRunningTime="2026-03-12 13:20:02.250087831 +0000 UTC m=+620.698783267" Mar 12 13:20:03 crc kubenswrapper[4778]: I0312 13:20:03.242445 4778 generic.go:334] "Generic (PLEG): container finished" podID="4c617404-7840-495c-80da-593af33f77d6" containerID="97b3a747ac158c0518500113b5af025bff04e06faaee081df03d1a06860f190f" exitCode=0 Mar 12 13:20:03 crc kubenswrapper[4778]: I0312 13:20:03.242542 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555360-vwflx" event={"ID":"4c617404-7840-495c-80da-593af33f77d6","Type":"ContainerDied","Data":"97b3a747ac158c0518500113b5af025bff04e06faaee081df03d1a06860f190f"} Mar 12 13:20:04 crc kubenswrapper[4778]: I0312 13:20:04.488747 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:04 crc kubenswrapper[4778]: I0312 13:20:04.585034 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzwxv\" (UniqueName: \"kubernetes.io/projected/4c617404-7840-495c-80da-593af33f77d6-kube-api-access-kzwxv\") pod \"4c617404-7840-495c-80da-593af33f77d6\" (UID: \"4c617404-7840-495c-80da-593af33f77d6\") " Mar 12 13:20:04 crc kubenswrapper[4778]: I0312 13:20:04.591698 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c617404-7840-495c-80da-593af33f77d6-kube-api-access-kzwxv" (OuterVolumeSpecName: "kube-api-access-kzwxv") pod "4c617404-7840-495c-80da-593af33f77d6" (UID: "4c617404-7840-495c-80da-593af33f77d6"). InnerVolumeSpecName "kube-api-access-kzwxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:20:04 crc kubenswrapper[4778]: I0312 13:20:04.686364 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzwxv\" (UniqueName: \"kubernetes.io/projected/4c617404-7840-495c-80da-593af33f77d6-kube-api-access-kzwxv\") on node \"crc\" DevicePath \"\"" Mar 12 13:20:05 crc kubenswrapper[4778]: I0312 13:20:05.256695 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555360-vwflx" event={"ID":"4c617404-7840-495c-80da-593af33f77d6","Type":"ContainerDied","Data":"0351e8b2cc8d22e69dee8334ff2e34916d4f30011467286d86f3b0031f613676"} Mar 12 13:20:05 crc kubenswrapper[4778]: I0312 13:20:05.256958 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0351e8b2cc8d22e69dee8334ff2e34916d4f30011467286d86f3b0031f613676" Mar 12 13:20:05 crc kubenswrapper[4778]: I0312 13:20:05.256779 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555360-vwflx" Mar 12 13:20:05 crc kubenswrapper[4778]: I0312 13:20:05.306946 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555354-n6zvc"] Mar 12 13:20:05 crc kubenswrapper[4778]: I0312 13:20:05.311727 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555354-n6zvc"] Mar 12 13:20:06 crc kubenswrapper[4778]: I0312 13:20:06.263814 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f91620d9-a95e-4e74-ab13-531d5e040b50" path="/var/lib/kubelet/pods/f91620d9-a95e-4e74-ab13-531d5e040b50/volumes" Mar 12 13:20:46 crc kubenswrapper[4778]: I0312 13:20:46.602740 4778 scope.go:117] "RemoveContainer" containerID="d6a4e00222817c0335bb85eb95073d869a129a695fed4bc12743392acf13e251" Mar 12 13:20:46 crc kubenswrapper[4778]: I0312 13:20:46.633549 4778 scope.go:117] "RemoveContainer" containerID="6a586e8ffe815ea410f687edd18208ce93300b26a8a15a7f7b6bd8396c76c788" Mar 12 13:20:46 crc kubenswrapper[4778]: I0312 13:20:46.664713 4778 scope.go:117] "RemoveContainer" containerID="1e77f31cb8ac97bbace99ce9835f811074e891b28dabf061e7039bfab7607d57" Mar 12 13:20:46 crc kubenswrapper[4778]: I0312 13:20:46.681801 4778 scope.go:117] "RemoveContainer" containerID="e6857324d1a49d08837ab795e083cf8ed33ad61f45f62f385bd7494ef38b2514" Mar 12 13:21:58 crc kubenswrapper[4778]: I0312 13:21:58.558485 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:21:58 crc kubenswrapper[4778]: I0312 13:21:58.559457 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.140486 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555362-hlj7f"] Mar 12 13:22:00 crc kubenswrapper[4778]: E0312 13:22:00.140799 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c617404-7840-495c-80da-593af33f77d6" containerName="oc" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.140815 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c617404-7840-495c-80da-593af33f77d6" containerName="oc" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.140917 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c617404-7840-495c-80da-593af33f77d6" containerName="oc" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.141467 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.144619 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.144806 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.147671 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555362-hlj7f"] Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.148557 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.211715 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqtjq\" (UniqueName: \"kubernetes.io/projected/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b-kube-api-access-dqtjq\") pod \"auto-csr-approver-29555362-hlj7f\" (UID: \"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b\") " pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.313667 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqtjq\" (UniqueName: \"kubernetes.io/projected/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b-kube-api-access-dqtjq\") pod \"auto-csr-approver-29555362-hlj7f\" (UID: \"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b\") " pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.336629 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqtjq\" (UniqueName: \"kubernetes.io/projected/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b-kube-api-access-dqtjq\") pod \"auto-csr-approver-29555362-hlj7f\" (UID: \"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b\") " pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.456862 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:00 crc kubenswrapper[4778]: I0312 13:22:00.886320 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555362-hlj7f"] Mar 12 13:22:00 crc kubenswrapper[4778]: W0312 13:22:00.900063 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9da11ea2_3173_4f25_8f0e_3ccc5a0ca18b.slice/crio-4cadda00557bd5590431db7bfe006b025eff40be2272cd2a46765ffa91761117 WatchSource:0}: Error finding container 4cadda00557bd5590431db7bfe006b025eff40be2272cd2a46765ffa91761117: Status 404 returned error can't find the container with id 4cadda00557bd5590431db7bfe006b025eff40be2272cd2a46765ffa91761117 Mar 12 13:22:01 crc kubenswrapper[4778]: I0312 13:22:01.315780 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" event={"ID":"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b","Type":"ContainerStarted","Data":"4cadda00557bd5590431db7bfe006b025eff40be2272cd2a46765ffa91761117"} Mar 12 13:22:03 crc kubenswrapper[4778]: I0312 13:22:03.331411 4778 generic.go:334] "Generic (PLEG): container finished" podID="9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b" containerID="1c6932f83080c12204b2bc10f63ca97fbee0fb238358dc69be9a27d4fc46a8a5" exitCode=0 Mar 12 13:22:03 crc kubenswrapper[4778]: I0312 13:22:03.331489 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" event={"ID":"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b","Type":"ContainerDied","Data":"1c6932f83080c12204b2bc10f63ca97fbee0fb238358dc69be9a27d4fc46a8a5"} Mar 12 13:22:04 crc kubenswrapper[4778]: I0312 13:22:04.591734 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:04 crc kubenswrapper[4778]: I0312 13:22:04.685038 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqtjq\" (UniqueName: \"kubernetes.io/projected/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b-kube-api-access-dqtjq\") pod \"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b\" (UID: \"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b\") " Mar 12 13:22:04 crc kubenswrapper[4778]: I0312 13:22:04.692080 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b-kube-api-access-dqtjq" (OuterVolumeSpecName: "kube-api-access-dqtjq") pod "9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b" (UID: "9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b"). InnerVolumeSpecName "kube-api-access-dqtjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:22:04 crc kubenswrapper[4778]: I0312 13:22:04.786001 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqtjq\" (UniqueName: \"kubernetes.io/projected/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b-kube-api-access-dqtjq\") on node \"crc\" DevicePath \"\"" Mar 12 13:22:05 crc kubenswrapper[4778]: I0312 13:22:05.342773 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" event={"ID":"9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b","Type":"ContainerDied","Data":"4cadda00557bd5590431db7bfe006b025eff40be2272cd2a46765ffa91761117"} Mar 12 13:22:05 crc kubenswrapper[4778]: I0312 13:22:05.342810 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555362-hlj7f" Mar 12 13:22:05 crc kubenswrapper[4778]: I0312 13:22:05.342821 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cadda00557bd5590431db7bfe006b025eff40be2272cd2a46765ffa91761117" Mar 12 13:22:05 crc kubenswrapper[4778]: I0312 13:22:05.647853 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555356-cdmcz"] Mar 12 13:22:05 crc kubenswrapper[4778]: I0312 13:22:05.651001 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555356-cdmcz"] Mar 12 13:22:06 crc kubenswrapper[4778]: I0312 13:22:06.266690 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c792e81a-8273-49a7-be95-c8c19cd2785b" path="/var/lib/kubelet/pods/c792e81a-8273-49a7-be95-c8c19cd2785b/volumes" Mar 12 13:22:28 crc kubenswrapper[4778]: I0312 13:22:28.557917 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:22:28 crc kubenswrapper[4778]: I0312 13:22:28.558412 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:22:46 crc kubenswrapper[4778]: I0312 13:22:46.755456 4778 scope.go:117] "RemoveContainer" containerID="b6d55e4553c4a90b5714d39c88d9e361c3f3109a89cdbda1980233a5b1fade38" Mar 12 13:22:58 crc kubenswrapper[4778]: I0312 13:22:58.558374 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:22:58 crc kubenswrapper[4778]: I0312 13:22:58.559247 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:22:58 crc kubenswrapper[4778]: I0312 13:22:58.559329 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:22:58 crc kubenswrapper[4778]: I0312 13:22:58.560062 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfcc37339849724c4aacca3262255dd43897a2284c2172380a90cc97f52e3a46"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:22:58 crc kubenswrapper[4778]: I0312 13:22:58.560151 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://dfcc37339849724c4aacca3262255dd43897a2284c2172380a90cc97f52e3a46" gracePeriod=600 Mar 12 13:22:59 crc kubenswrapper[4778]: I0312 13:22:59.689390 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="dfcc37339849724c4aacca3262255dd43897a2284c2172380a90cc97f52e3a46" exitCode=0 Mar 12 13:22:59 crc kubenswrapper[4778]: I0312 13:22:59.689464 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"dfcc37339849724c4aacca3262255dd43897a2284c2172380a90cc97f52e3a46"} Mar 12 13:22:59 crc kubenswrapper[4778]: I0312 13:22:59.689807 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"b65e287d42eea6146877a35b0789c26ac0ef9f5d251a760b59f08b3fef055d65"} Mar 12 13:22:59 crc kubenswrapper[4778]: I0312 13:22:59.689830 4778 scope.go:117] "RemoveContainer" containerID="e50690e6aff1fd408e6201d1eee1240e692ce04bc21873dbbe85a5f2d638d704" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.877216 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g"] Mar 12 13:23:04 crc kubenswrapper[4778]: E0312 13:23:04.877984 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b" containerName="oc" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.878001 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b" containerName="oc" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.878129 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b" containerName="oc" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.878586 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.881240 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.881686 4778 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-x4ptm" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.881995 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.887074 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-2774s"] Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.887985 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-2774s" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.889895 4778 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-r9jbf" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.908480 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-2774s"] Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.913500 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g"] Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.933424 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-ffh2x"] Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.935642 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.937666 4778 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6zmz6" Mar 12 13:23:04 crc kubenswrapper[4778]: I0312 13:23:04.942193 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-ffh2x"] Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.032666 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmkwl\" (UniqueName: \"kubernetes.io/projected/804d0b09-6fab-4277-936a-5e0324d76b3e-kube-api-access-xmkwl\") pod \"cert-manager-cainjector-cf98fcc89-jxs4g\" (UID: \"804d0b09-6fab-4277-936a-5e0324d76b3e\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.032741 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-668fd\" (UniqueName: \"kubernetes.io/projected/45da07c5-bccb-4433-aa38-d9d2894f1b09-kube-api-access-668fd\") pod \"cert-manager-webhook-687f57d79b-ffh2x\" (UID: \"45da07c5-bccb-4433-aa38-d9d2894f1b09\") " pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.032777 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m52wz\" (UniqueName: \"kubernetes.io/projected/92b29110-f478-42b5-9a5f-c9330a3973b2-kube-api-access-m52wz\") pod \"cert-manager-858654f9db-2774s\" (UID: \"92b29110-f478-42b5-9a5f-c9330a3973b2\") " pod="cert-manager/cert-manager-858654f9db-2774s" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.134312 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmkwl\" (UniqueName: \"kubernetes.io/projected/804d0b09-6fab-4277-936a-5e0324d76b3e-kube-api-access-xmkwl\") pod \"cert-manager-cainjector-cf98fcc89-jxs4g\" (UID: \"804d0b09-6fab-4277-936a-5e0324d76b3e\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.134388 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-668fd\" (UniqueName: \"kubernetes.io/projected/45da07c5-bccb-4433-aa38-d9d2894f1b09-kube-api-access-668fd\") pod \"cert-manager-webhook-687f57d79b-ffh2x\" (UID: \"45da07c5-bccb-4433-aa38-d9d2894f1b09\") " pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.134419 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m52wz\" (UniqueName: \"kubernetes.io/projected/92b29110-f478-42b5-9a5f-c9330a3973b2-kube-api-access-m52wz\") pod \"cert-manager-858654f9db-2774s\" (UID: \"92b29110-f478-42b5-9a5f-c9330a3973b2\") " pod="cert-manager/cert-manager-858654f9db-2774s" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.153157 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m52wz\" (UniqueName: \"kubernetes.io/projected/92b29110-f478-42b5-9a5f-c9330a3973b2-kube-api-access-m52wz\") pod \"cert-manager-858654f9db-2774s\" (UID: \"92b29110-f478-42b5-9a5f-c9330a3973b2\") " pod="cert-manager/cert-manager-858654f9db-2774s" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.153735 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmkwl\" (UniqueName: \"kubernetes.io/projected/804d0b09-6fab-4277-936a-5e0324d76b3e-kube-api-access-xmkwl\") pod \"cert-manager-cainjector-cf98fcc89-jxs4g\" (UID: \"804d0b09-6fab-4277-936a-5e0324d76b3e\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.158435 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-668fd\" (UniqueName: \"kubernetes.io/projected/45da07c5-bccb-4433-aa38-d9d2894f1b09-kube-api-access-668fd\") pod \"cert-manager-webhook-687f57d79b-ffh2x\" (UID: \"45da07c5-bccb-4433-aa38-d9d2894f1b09\") " pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.191702 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.209546 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-2774s" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.252442 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.431441 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g"] Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.444745 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.468303 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-2774s"] Mar 12 13:23:05 crc kubenswrapper[4778]: W0312 13:23:05.472393 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92b29110_f478_42b5_9a5f_c9330a3973b2.slice/crio-7649f3a93dee22f65d3b47331535f4c516f53f9ee9e71de7aa8bd889a816a4b2 WatchSource:0}: Error finding container 7649f3a93dee22f65d3b47331535f4c516f53f9ee9e71de7aa8bd889a816a4b2: Status 404 returned error can't find the container with id 7649f3a93dee22f65d3b47331535f4c516f53f9ee9e71de7aa8bd889a816a4b2 Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.520824 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-ffh2x"] Mar 12 13:23:05 crc kubenswrapper[4778]: W0312 13:23:05.532385 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45da07c5_bccb_4433_aa38_d9d2894f1b09.slice/crio-02763cee9bdd368e2df4e6001abd9f67d450f69519235734d269a72af5e041e2 WatchSource:0}: Error finding container 02763cee9bdd368e2df4e6001abd9f67d450f69519235734d269a72af5e041e2: Status 404 returned error can't find the container with id 02763cee9bdd368e2df4e6001abd9f67d450f69519235734d269a72af5e041e2 Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.724232 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-2774s" event={"ID":"92b29110-f478-42b5-9a5f-c9330a3973b2","Type":"ContainerStarted","Data":"7649f3a93dee22f65d3b47331535f4c516f53f9ee9e71de7aa8bd889a816a4b2"} Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.725442 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" event={"ID":"45da07c5-bccb-4433-aa38-d9d2894f1b09","Type":"ContainerStarted","Data":"02763cee9bdd368e2df4e6001abd9f67d450f69519235734d269a72af5e041e2"} Mar 12 13:23:05 crc kubenswrapper[4778]: I0312 13:23:05.726834 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" event={"ID":"804d0b09-6fab-4277-936a-5e0324d76b3e","Type":"ContainerStarted","Data":"10510c6ad52b31596235e39038f9df629faf2cb933dd3c897d53107b62a0addb"} Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.779424 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-2774s" event={"ID":"92b29110-f478-42b5-9a5f-c9330a3973b2","Type":"ContainerStarted","Data":"11efd69e962344516667a5b7b415e9dcfff2f837949701433b301278c93e8b43"} Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.781956 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" event={"ID":"45da07c5-bccb-4433-aa38-d9d2894f1b09","Type":"ContainerStarted","Data":"d813356578c5e1f7bab8605e1f5ad4fb6a22081d657c6f914c75bb9d59e890ef"} Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.782079 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.783728 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" event={"ID":"804d0b09-6fab-4277-936a-5e0324d76b3e","Type":"ContainerStarted","Data":"725cb330b7b3d6799189ae3a295da601d5d574bfb6cffd99984951cfd40f6425"} Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.798347 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-2774s" podStartSLOduration=2.624663656 podStartE2EDuration="8.798309642s" podCreationTimestamp="2026-03-12 13:23:04 +0000 UTC" firstStartedPulling="2026-03-12 13:23:05.475227653 +0000 UTC m=+803.923923049" lastFinishedPulling="2026-03-12 13:23:11.648873639 +0000 UTC m=+810.097569035" observedRunningTime="2026-03-12 13:23:12.795417129 +0000 UTC m=+811.244112555" watchObservedRunningTime="2026-03-12 13:23:12.798309642 +0000 UTC m=+811.247005048" Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.814540 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" podStartSLOduration=2.56655417 podStartE2EDuration="8.814501966s" podCreationTimestamp="2026-03-12 13:23:04 +0000 UTC" firstStartedPulling="2026-03-12 13:23:05.535444108 +0000 UTC m=+803.984139514" lastFinishedPulling="2026-03-12 13:23:11.783391914 +0000 UTC m=+810.232087310" observedRunningTime="2026-03-12 13:23:12.810839521 +0000 UTC m=+811.259534917" watchObservedRunningTime="2026-03-12 13:23:12.814501966 +0000 UTC m=+811.263197362" Mar 12 13:23:12 crc kubenswrapper[4778]: I0312 13:23:12.843791 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jxs4g" podStartSLOduration=2.632979375 podStartE2EDuration="8.843772055s" podCreationTimestamp="2026-03-12 13:23:04 +0000 UTC" firstStartedPulling="2026-03-12 13:23:05.443829903 +0000 UTC m=+803.892525299" lastFinishedPulling="2026-03-12 13:23:11.654622583 +0000 UTC m=+810.103317979" observedRunningTime="2026-03-12 13:23:12.842630852 +0000 UTC m=+811.291326248" watchObservedRunningTime="2026-03-12 13:23:12.843772055 +0000 UTC m=+811.292467461" Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.732152 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8bcc9"] Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.733898 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-controller" containerID="cri-o://b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.734122 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="northd" containerID="cri-o://2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.734223 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.734094 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="sbdb" containerID="cri-o://6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.734273 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-acl-logging" containerID="cri-o://1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.733990 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="nbdb" containerID="cri-o://78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.734107 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-node" containerID="cri-o://8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" gracePeriod=30 Mar 12 13:23:14 crc kubenswrapper[4778]: I0312 13:23:14.779349 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" containerID="cri-o://9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" gracePeriod=30 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.549092 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/3.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.552704 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovn-acl-logging/0.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.553519 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovn-controller/0.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.554225 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.618989 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b8gxm"] Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619197 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-node" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619209 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-node" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619221 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kubecfg-setup" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619262 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kubecfg-setup" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619271 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-acl-logging" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619276 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-acl-logging" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619287 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="nbdb" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619292 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="nbdb" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619299 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619305 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619313 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619318 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619328 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="sbdb" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619334 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="sbdb" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619344 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619349 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619355 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619361 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619368 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619373 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619382 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="northd" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619388 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="northd" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619393 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619399 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: E0312 13:23:15.619408 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619414 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619500 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-acl-logging" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619512 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619520 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="nbdb" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619530 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-node" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619536 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="northd" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619543 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="sbdb" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619550 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619558 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovn-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619565 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619571 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619577 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.619737 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerName="ovnkube-controller" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.621350 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680761 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-log-socket\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680794 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-netns\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680825 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-config\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680849 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-slash\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680874 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-kubelet\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680914 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-etc-openvswitch\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680933 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-ovn-kubernetes\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680947 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-ovn\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680978 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovn-node-metrics-cert\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.680998 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-systemd\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681011 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-node-log\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681029 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681048 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-script-lib\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681063 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-var-lib-openvswitch\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681081 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-openvswitch\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681100 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-bin\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681113 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-netd\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681130 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-env-overrides\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681150 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-schvw\" (UniqueName: \"kubernetes.io/projected/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-kube-api-access-schvw\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681166 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-systemd-units\") pod \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\" (UID: \"65cd795e-eb6e-4995-a4c1-9dea6f425ac5\") " Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681378 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681408 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-log-socket" (OuterVolumeSpecName: "log-socket") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681424 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681806 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681831 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-slash" (OuterVolumeSpecName: "host-slash") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681847 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681863 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681878 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.681895 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.682663 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.682701 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-node-log" (OuterVolumeSpecName: "node-log") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.682726 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.682754 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.682820 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.682850 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.683002 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.683307 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.687340 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-kube-api-access-schvw" (OuterVolumeSpecName: "kube-api-access-schvw") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "kube-api-access-schvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.688061 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.696728 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "65cd795e-eb6e-4995-a4c1-9dea6f425ac5" (UID: "65cd795e-eb6e-4995-a4c1-9dea6f425ac5"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.782273 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-log-socket\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.783522 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovn-node-metrics-cert\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.783677 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-env-overrides\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.783803 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfssb\" (UniqueName: \"kubernetes.io/projected/1664bb45-1e97-4371-9dbc-5e27e778ee0b-kube-api-access-cfssb\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.783931 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-ovn\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784040 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-cni-bin\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784175 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784294 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-cni-netd\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784441 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovnkube-script-lib\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784558 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-slash\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784679 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-systemd-units\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.784814 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-var-lib-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785092 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-run-netns\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785232 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-systemd\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785355 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovnkube-config\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785561 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785671 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-node-log\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785794 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-etc-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.785891 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786019 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-kubelet\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786234 4778 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786323 4778 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786398 4778 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786471 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-schvw\" (UniqueName: \"kubernetes.io/projected/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-kube-api-access-schvw\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786544 4778 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786622 4778 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-log-socket\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786688 4778 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786752 4778 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786854 4778 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-slash\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786930 4778 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.786998 4778 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787073 4778 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787177 4778 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787270 4778 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787374 4778 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787439 4778 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-node-log\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787494 4778 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787554 4778 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787607 4778 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.787661 4778 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/65cd795e-eb6e-4995-a4c1-9dea6f425ac5-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.804076 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fhcz6_1e7037a8-a966-4df0-9f94-fe2dd3e2de6e/kube-multus/1.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.804762 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fhcz6_1e7037a8-a966-4df0-9f94-fe2dd3e2de6e/kube-multus/0.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.804897 4778 generic.go:334] "Generic (PLEG): container finished" podID="1e7037a8-a966-4df0-9f94-fe2dd3e2de6e" containerID="44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e" exitCode=2 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.805040 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fhcz6" event={"ID":"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e","Type":"ContainerDied","Data":"44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.805212 4778 scope.go:117] "RemoveContainer" containerID="5da98f94c85e3a8cd05c447fb097a078968eea25419a2b22f8abe956ef1dbaac" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.805917 4778 scope.go:117] "RemoveContainer" containerID="44a3c76b2249ac9c24848e6b3a9fc08aef2d2bca3d170ce28b0f9384e3a8271e" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.811181 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovnkube-controller/3.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.815138 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovn-acl-logging/0.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.815776 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8bcc9_65cd795e-eb6e-4995-a4c1-9dea6f425ac5/ovn-controller/0.log" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816227 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" exitCode=0 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816251 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" exitCode=0 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816260 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" exitCode=0 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816269 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" exitCode=0 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816277 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" exitCode=0 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816285 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" exitCode=0 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816293 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" exitCode=143 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816254 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816380 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816300 4778 generic.go:334] "Generic (PLEG): container finished" podID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" exitCode=143 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816408 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816424 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816471 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816485 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816498 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816511 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816545 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816554 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816561 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816568 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816575 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816583 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816590 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816597 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816632 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816647 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816655 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816664 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816671 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816678 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816710 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816719 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816726 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816734 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816741 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816752 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816764 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816799 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816807 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816814 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816821 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816828 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816836 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816843 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816876 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816884 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816895 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" event={"ID":"65cd795e-eb6e-4995-a4c1-9dea6f425ac5","Type":"ContainerDied","Data":"591e87d9e47004fc9c6fc7b24484cec488177d8e0820b4787eb9618d9e5051df"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816907 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816915 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816923 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816954 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816964 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816971 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816979 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.816986 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.817028 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.817036 4778 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.818328 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8bcc9" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.863227 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8bcc9"] Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.869132 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8bcc9"] Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889412 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-var-lib-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889523 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-run-netns\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889553 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-systemd\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889573 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovnkube-config\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889600 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889605 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-var-lib-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889622 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-node-log\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889677 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-run-netns\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889713 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-systemd\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889691 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-node-log\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889713 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-etc-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889742 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-etc-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889767 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889794 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889770 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889846 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-kubelet\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889890 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-log-socket\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889916 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovn-node-metrics-cert\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889939 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-log-socket\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889945 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-env-overrides\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890008 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfssb\" (UniqueName: \"kubernetes.io/projected/1664bb45-1e97-4371-9dbc-5e27e778ee0b-kube-api-access-cfssb\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890041 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-ovn\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890073 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-cni-bin\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890122 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890159 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-cni-netd\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890226 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovnkube-script-lib\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890250 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-slash\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890272 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-systemd-units\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890369 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-systemd-units\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890509 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-env-overrides\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890697 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-ovn\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890732 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-cni-bin\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890760 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-run-openvswitch\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890789 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-cni-netd\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.890930 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-slash\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.889913 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1664bb45-1e97-4371-9dbc-5e27e778ee0b-host-kubelet\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.891455 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovnkube-script-lib\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.895748 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovn-node-metrics-cert\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.896539 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1664bb45-1e97-4371-9dbc-5e27e778ee0b-ovnkube-config\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.909170 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfssb\" (UniqueName: \"kubernetes.io/projected/1664bb45-1e97-4371-9dbc-5e27e778ee0b-kube-api-access-cfssb\") pod \"ovnkube-node-b8gxm\" (UID: \"1664bb45-1e97-4371-9dbc-5e27e778ee0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.940100 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:15 crc kubenswrapper[4778]: W0312 13:23:15.957888 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1664bb45_1e97_4371_9dbc_5e27e778ee0b.slice/crio-01d8ef80533ea4fb98aabf7de84e5f17da4cb625b7e47d4badaaa85038ae83f5 WatchSource:0}: Error finding container 01d8ef80533ea4fb98aabf7de84e5f17da4cb625b7e47d4badaaa85038ae83f5: Status 404 returned error can't find the container with id 01d8ef80533ea4fb98aabf7de84e5f17da4cb625b7e47d4badaaa85038ae83f5 Mar 12 13:23:15 crc kubenswrapper[4778]: I0312 13:23:15.992896 4778 scope.go:117] "RemoveContainer" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.010747 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.026394 4778 scope.go:117] "RemoveContainer" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.045002 4778 scope.go:117] "RemoveContainer" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.057847 4778 scope.go:117] "RemoveContainer" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.070895 4778 scope.go:117] "RemoveContainer" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.097976 4778 scope.go:117] "RemoveContainer" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.111228 4778 scope.go:117] "RemoveContainer" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.126507 4778 scope.go:117] "RemoveContainer" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.138130 4778 scope.go:117] "RemoveContainer" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.150704 4778 scope.go:117] "RemoveContainer" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.151203 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": container with ID starting with 9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d not found: ID does not exist" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.151250 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} err="failed to get container status \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": rpc error: code = NotFound desc = could not find container \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": container with ID starting with 9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.151273 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.151590 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": container with ID starting with 5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001 not found: ID does not exist" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.151638 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} err="failed to get container status \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": rpc error: code = NotFound desc = could not find container \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": container with ID starting with 5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.151674 4778 scope.go:117] "RemoveContainer" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.152039 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": container with ID starting with 6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d not found: ID does not exist" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.152073 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} err="failed to get container status \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": rpc error: code = NotFound desc = could not find container \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": container with ID starting with 6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.152092 4778 scope.go:117] "RemoveContainer" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.152362 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": container with ID starting with 78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa not found: ID does not exist" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.152388 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} err="failed to get container status \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": rpc error: code = NotFound desc = could not find container \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": container with ID starting with 78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.152405 4778 scope.go:117] "RemoveContainer" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.152760 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": container with ID starting with 2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483 not found: ID does not exist" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.152786 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} err="failed to get container status \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": rpc error: code = NotFound desc = could not find container \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": container with ID starting with 2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.152805 4778 scope.go:117] "RemoveContainer" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.153065 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": container with ID starting with 1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e not found: ID does not exist" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153098 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} err="failed to get container status \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": rpc error: code = NotFound desc = could not find container \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": container with ID starting with 1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153118 4778 scope.go:117] "RemoveContainer" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.153364 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": container with ID starting with 8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4 not found: ID does not exist" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153387 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} err="failed to get container status \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": rpc error: code = NotFound desc = could not find container \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": container with ID starting with 8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153405 4778 scope.go:117] "RemoveContainer" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.153652 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": container with ID starting with 1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6 not found: ID does not exist" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153682 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} err="failed to get container status \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": rpc error: code = NotFound desc = could not find container \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": container with ID starting with 1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153701 4778 scope.go:117] "RemoveContainer" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.153958 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": container with ID starting with b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500 not found: ID does not exist" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153981 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} err="failed to get container status \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": rpc error: code = NotFound desc = could not find container \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": container with ID starting with b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.153995 4778 scope.go:117] "RemoveContainer" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" Mar 12 13:23:16 crc kubenswrapper[4778]: E0312 13:23:16.154229 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": container with ID starting with ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e not found: ID does not exist" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.154254 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} err="failed to get container status \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": rpc error: code = NotFound desc = could not find container \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": container with ID starting with ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.154266 4778 scope.go:117] "RemoveContainer" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.154665 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} err="failed to get container status \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": rpc error: code = NotFound desc = could not find container \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": container with ID starting with 9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.154691 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.154923 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} err="failed to get container status \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": rpc error: code = NotFound desc = could not find container \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": container with ID starting with 5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.154949 4778 scope.go:117] "RemoveContainer" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.155195 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} err="failed to get container status \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": rpc error: code = NotFound desc = could not find container \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": container with ID starting with 6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.155221 4778 scope.go:117] "RemoveContainer" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.155456 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} err="failed to get container status \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": rpc error: code = NotFound desc = could not find container \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": container with ID starting with 78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.155480 4778 scope.go:117] "RemoveContainer" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.155787 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} err="failed to get container status \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": rpc error: code = NotFound desc = could not find container \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": container with ID starting with 2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.155807 4778 scope.go:117] "RemoveContainer" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.156046 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} err="failed to get container status \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": rpc error: code = NotFound desc = could not find container \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": container with ID starting with 1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.156061 4778 scope.go:117] "RemoveContainer" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.156306 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} err="failed to get container status \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": rpc error: code = NotFound desc = could not find container \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": container with ID starting with 8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.156327 4778 scope.go:117] "RemoveContainer" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.156625 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} err="failed to get container status \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": rpc error: code = NotFound desc = could not find container \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": container with ID starting with 1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.156647 4778 scope.go:117] "RemoveContainer" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.157320 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} err="failed to get container status \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": rpc error: code = NotFound desc = could not find container \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": container with ID starting with b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.157343 4778 scope.go:117] "RemoveContainer" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.157680 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} err="failed to get container status \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": rpc error: code = NotFound desc = could not find container \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": container with ID starting with ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.157708 4778 scope.go:117] "RemoveContainer" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158045 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} err="failed to get container status \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": rpc error: code = NotFound desc = could not find container \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": container with ID starting with 9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158066 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158300 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} err="failed to get container status \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": rpc error: code = NotFound desc = could not find container \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": container with ID starting with 5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158334 4778 scope.go:117] "RemoveContainer" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158663 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} err="failed to get container status \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": rpc error: code = NotFound desc = could not find container \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": container with ID starting with 6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158695 4778 scope.go:117] "RemoveContainer" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.158997 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} err="failed to get container status \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": rpc error: code = NotFound desc = could not find container \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": container with ID starting with 78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159025 4778 scope.go:117] "RemoveContainer" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159272 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} err="failed to get container status \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": rpc error: code = NotFound desc = could not find container \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": container with ID starting with 2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159296 4778 scope.go:117] "RemoveContainer" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159501 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} err="failed to get container status \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": rpc error: code = NotFound desc = could not find container \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": container with ID starting with 1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159556 4778 scope.go:117] "RemoveContainer" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159767 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} err="failed to get container status \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": rpc error: code = NotFound desc = could not find container \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": container with ID starting with 8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159787 4778 scope.go:117] "RemoveContainer" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.159994 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} err="failed to get container status \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": rpc error: code = NotFound desc = could not find container \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": container with ID starting with 1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.160016 4778 scope.go:117] "RemoveContainer" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.160428 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} err="failed to get container status \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": rpc error: code = NotFound desc = could not find container \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": container with ID starting with b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.160467 4778 scope.go:117] "RemoveContainer" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.160782 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} err="failed to get container status \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": rpc error: code = NotFound desc = could not find container \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": container with ID starting with ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.160804 4778 scope.go:117] "RemoveContainer" containerID="9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.161120 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d"} err="failed to get container status \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": rpc error: code = NotFound desc = could not find container \"9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d\": container with ID starting with 9afb5c8d21c64a6b41dbded768a82ec790fb6f2f6a21efa119251504eb0c3a8d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.161147 4778 scope.go:117] "RemoveContainer" containerID="5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.161470 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001"} err="failed to get container status \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": rpc error: code = NotFound desc = could not find container \"5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001\": container with ID starting with 5d6da6dba0e8cadf9b1073620c4856adeb6b776ae3757d420c016d25b4f98001 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.161493 4778 scope.go:117] "RemoveContainer" containerID="6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.161713 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d"} err="failed to get container status \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": rpc error: code = NotFound desc = could not find container \"6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d\": container with ID starting with 6bc4107e3fb5708a2acf2664ce876af4c682377a9b4ee393230b78d5b021552d not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.161733 4778 scope.go:117] "RemoveContainer" containerID="78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162001 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa"} err="failed to get container status \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": rpc error: code = NotFound desc = could not find container \"78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa\": container with ID starting with 78da788cc1d96e866afcf18edff24c064972e022b1b4c3f5bec3175da5e989fa not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162024 4778 scope.go:117] "RemoveContainer" containerID="2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162323 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483"} err="failed to get container status \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": rpc error: code = NotFound desc = could not find container \"2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483\": container with ID starting with 2818aadc56c24df41309aa63fddf44dac870f041f206d59db4d8b8f88f728483 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162340 4778 scope.go:117] "RemoveContainer" containerID="1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162529 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e"} err="failed to get container status \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": rpc error: code = NotFound desc = could not find container \"1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e\": container with ID starting with 1325a4d5784843f5ef2dc629d3410d154ea07de5cd70cfae54d7111fe3e1ea3e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162551 4778 scope.go:117] "RemoveContainer" containerID="8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162842 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4"} err="failed to get container status \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": rpc error: code = NotFound desc = could not find container \"8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4\": container with ID starting with 8f7250fa81a99a607a87f5f2fe4a85fcc2afbf7bfc12845bec4b5cbbed7784c4 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.162861 4778 scope.go:117] "RemoveContainer" containerID="1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.163240 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6"} err="failed to get container status \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": rpc error: code = NotFound desc = could not find container \"1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6\": container with ID starting with 1d8f2c481e6f3f8845b5e195a7fde8ae6415d25fb1701c25d93be7af1f4ef8f6 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.163311 4778 scope.go:117] "RemoveContainer" containerID="b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.163656 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500"} err="failed to get container status \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": rpc error: code = NotFound desc = could not find container \"b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500\": container with ID starting with b22f94192e4eee991a699c32f338a8d452d8fc0ce5dfa1197694a237697c4500 not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.163676 4778 scope.go:117] "RemoveContainer" containerID="ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.163943 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e"} err="failed to get container status \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": rpc error: code = NotFound desc = could not find container \"ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e\": container with ID starting with ce861cdc0bd23c8ce1b2989859dc05ceab2b6af01ac92029c419edc7f58c2b1e not found: ID does not exist" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.284465 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65cd795e-eb6e-4995-a4c1-9dea6f425ac5" path="/var/lib/kubelet/pods/65cd795e-eb6e-4995-a4c1-9dea6f425ac5/volumes" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.824754 4778 generic.go:334] "Generic (PLEG): container finished" podID="1664bb45-1e97-4371-9dbc-5e27e778ee0b" containerID="a1d264705e7133e0eb93e53519cdbe2e459c8f6e5a407d4657ed576e12d2b397" exitCode=0 Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.824857 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerDied","Data":"a1d264705e7133e0eb93e53519cdbe2e459c8f6e5a407d4657ed576e12d2b397"} Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.824917 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"01d8ef80533ea4fb98aabf7de84e5f17da4cb625b7e47d4badaaa85038ae83f5"} Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.829506 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fhcz6_1e7037a8-a966-4df0-9f94-fe2dd3e2de6e/kube-multus/1.log" Mar 12 13:23:16 crc kubenswrapper[4778]: I0312 13:23:16.829644 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fhcz6" event={"ID":"1e7037a8-a966-4df0-9f94-fe2dd3e2de6e","Type":"ContainerStarted","Data":"dbb67014f504889121924b5a7d01a6364d0717cc22842ad77b3827450b0bcf2c"} Mar 12 13:23:17 crc kubenswrapper[4778]: I0312 13:23:17.841157 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"9f6c54c39750a31fad04289db347d8fd7e01a78681b55a1b3b20c443d64600fc"} Mar 12 13:23:17 crc kubenswrapper[4778]: I0312 13:23:17.841457 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"8106201d2dce337561821161936bc1e2c5d1cb6f072033e304e2788a17f59091"} Mar 12 13:23:17 crc kubenswrapper[4778]: I0312 13:23:17.841469 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"02dbf475c36f92dbffe0ed598faf01abcbb066bee25f8b9b7158434cfc9cdcf9"} Mar 12 13:23:17 crc kubenswrapper[4778]: I0312 13:23:17.841478 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"d91aabc97edc08ee6b0fedb422db1f5f30b6a7c730554fa4fd2e70c02079cb28"} Mar 12 13:23:17 crc kubenswrapper[4778]: I0312 13:23:17.841486 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"3d68915fde1342954bdeedae820f9aca1d9e38fc407c53c550c6f5c1c2436f5a"} Mar 12 13:23:17 crc kubenswrapper[4778]: I0312 13:23:17.841494 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"c48ba3990b90a79802f5ed2b8c4baf9bdf992c1604fd9cb1292028cbb43d85ad"} Mar 12 13:23:20 crc kubenswrapper[4778]: I0312 13:23:20.260711 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-ffh2x" Mar 12 13:23:20 crc kubenswrapper[4778]: I0312 13:23:20.881037 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"4c7a2afa17744618a18e16336cd670680d0f5dc381f870cec07b0ad5d301079f"} Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.899728 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" event={"ID":"1664bb45-1e97-4371-9dbc-5e27e778ee0b","Type":"ContainerStarted","Data":"41c943533a50230bf7606d72941dd077da20bbd6ef0f0343292050596ef7a8b5"} Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.900334 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.900449 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.900519 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.936534 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" podStartSLOduration=7.9365080930000005 podStartE2EDuration="7.936508093s" podCreationTimestamp="2026-03-12 13:23:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:23:22.932870268 +0000 UTC m=+821.381565674" watchObservedRunningTime="2026-03-12 13:23:22.936508093 +0000 UTC m=+821.385203489" Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.938281 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:22 crc kubenswrapper[4778]: I0312 13:23:22.946128 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:23:45 crc kubenswrapper[4778]: I0312 13:23:45.961801 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b8gxm" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.143359 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555364-hrrdv"] Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.145290 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.148959 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555364-hrrdv"] Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.149797 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.150827 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.151800 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.223525 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmmcq\" (UniqueName: \"kubernetes.io/projected/c862c78c-5987-48cc-8b41-531755f319e9-kube-api-access-rmmcq\") pod \"auto-csr-approver-29555364-hrrdv\" (UID: \"c862c78c-5987-48cc-8b41-531755f319e9\") " pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.325070 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmmcq\" (UniqueName: \"kubernetes.io/projected/c862c78c-5987-48cc-8b41-531755f319e9-kube-api-access-rmmcq\") pod \"auto-csr-approver-29555364-hrrdv\" (UID: \"c862c78c-5987-48cc-8b41-531755f319e9\") " pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.352117 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmmcq\" (UniqueName: \"kubernetes.io/projected/c862c78c-5987-48cc-8b41-531755f319e9-kube-api-access-rmmcq\") pod \"auto-csr-approver-29555364-hrrdv\" (UID: \"c862c78c-5987-48cc-8b41-531755f319e9\") " pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.472510 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.657560 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555364-hrrdv"] Mar 12 13:24:00 crc kubenswrapper[4778]: I0312 13:24:00.662650 4778 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 12 13:24:01 crc kubenswrapper[4778]: I0312 13:24:01.121763 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" event={"ID":"c862c78c-5987-48cc-8b41-531755f319e9","Type":"ContainerStarted","Data":"45621410278cad64ed5807476fd5a76b1c6c29754bfb3cc45082b87d5cf1bb51"} Mar 12 13:24:03 crc kubenswrapper[4778]: I0312 13:24:03.140469 4778 generic.go:334] "Generic (PLEG): container finished" podID="c862c78c-5987-48cc-8b41-531755f319e9" containerID="a8f045f157371374b81f9a3098c61d715d2ce620fdfc3121b5f225672622998f" exitCode=0 Mar 12 13:24:03 crc kubenswrapper[4778]: I0312 13:24:03.140589 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" event={"ID":"c862c78c-5987-48cc-8b41-531755f319e9","Type":"ContainerDied","Data":"a8f045f157371374b81f9a3098c61d715d2ce620fdfc3121b5f225672622998f"} Mar 12 13:24:04 crc kubenswrapper[4778]: I0312 13:24:04.369883 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:04 crc kubenswrapper[4778]: I0312 13:24:04.472161 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmmcq\" (UniqueName: \"kubernetes.io/projected/c862c78c-5987-48cc-8b41-531755f319e9-kube-api-access-rmmcq\") pod \"c862c78c-5987-48cc-8b41-531755f319e9\" (UID: \"c862c78c-5987-48cc-8b41-531755f319e9\") " Mar 12 13:24:04 crc kubenswrapper[4778]: I0312 13:24:04.477144 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c862c78c-5987-48cc-8b41-531755f319e9-kube-api-access-rmmcq" (OuterVolumeSpecName: "kube-api-access-rmmcq") pod "c862c78c-5987-48cc-8b41-531755f319e9" (UID: "c862c78c-5987-48cc-8b41-531755f319e9"). InnerVolumeSpecName "kube-api-access-rmmcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:24:04 crc kubenswrapper[4778]: I0312 13:24:04.573647 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmmcq\" (UniqueName: \"kubernetes.io/projected/c862c78c-5987-48cc-8b41-531755f319e9-kube-api-access-rmmcq\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:05 crc kubenswrapper[4778]: I0312 13:24:05.161242 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" event={"ID":"c862c78c-5987-48cc-8b41-531755f319e9","Type":"ContainerDied","Data":"45621410278cad64ed5807476fd5a76b1c6c29754bfb3cc45082b87d5cf1bb51"} Mar 12 13:24:05 crc kubenswrapper[4778]: I0312 13:24:05.161581 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45621410278cad64ed5807476fd5a76b1c6c29754bfb3cc45082b87d5cf1bb51" Mar 12 13:24:05 crc kubenswrapper[4778]: I0312 13:24:05.161711 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555364-hrrdv" Mar 12 13:24:05 crc kubenswrapper[4778]: I0312 13:24:05.440830 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555358-txmvp"] Mar 12 13:24:05 crc kubenswrapper[4778]: I0312 13:24:05.444214 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555358-txmvp"] Mar 12 13:24:06 crc kubenswrapper[4778]: I0312 13:24:06.261227 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c6485d-2d53-47d9-866a-31eb90ac254e" path="/var/lib/kubelet/pods/61c6485d-2d53-47d9-866a-31eb90ac254e/volumes" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.521627 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd"] Mar 12 13:24:14 crc kubenswrapper[4778]: E0312 13:24:14.522131 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c862c78c-5987-48cc-8b41-531755f319e9" containerName="oc" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.522146 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c862c78c-5987-48cc-8b41-531755f319e9" containerName="oc" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.522280 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c862c78c-5987-48cc-8b41-531755f319e9" containerName="oc" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.522976 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.526260 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.532283 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd"] Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.612720 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.612802 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.612885 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm5dc\" (UniqueName: \"kubernetes.io/projected/cb93062b-8387-4eb4-8662-ecaf93146d85-kube-api-access-gm5dc\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.714697 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.714773 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.714860 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm5dc\" (UniqueName: \"kubernetes.io/projected/cb93062b-8387-4eb4-8662-ecaf93146d85-kube-api-access-gm5dc\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.715799 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.715911 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.750950 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm5dc\" (UniqueName: \"kubernetes.io/projected/cb93062b-8387-4eb4-8662-ecaf93146d85-kube-api-access-gm5dc\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:14 crc kubenswrapper[4778]: I0312 13:24:14.840305 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:15 crc kubenswrapper[4778]: I0312 13:24:15.340616 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd"] Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.222641 4778 generic.go:334] "Generic (PLEG): container finished" podID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerID="aeb9887345bc844496df7d0c5b6bf02eb4f105d1e813edce48f3ff227e3b96e2" exitCode=0 Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.222701 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" event={"ID":"cb93062b-8387-4eb4-8662-ecaf93146d85","Type":"ContainerDied","Data":"aeb9887345bc844496df7d0c5b6bf02eb4f105d1e813edce48f3ff227e3b96e2"} Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.222912 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" event={"ID":"cb93062b-8387-4eb4-8662-ecaf93146d85","Type":"ContainerStarted","Data":"3f052fe9978b090dcb12eafc251984a2e5ac0055088087a2227d82364c78a56d"} Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.749424 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tpgqw"] Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.750727 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.769271 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tpgqw"] Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.845791 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-catalog-content\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.845849 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99g7g\" (UniqueName: \"kubernetes.io/projected/b4c13351-b8fa-4224-a09b-942200d398b1-kube-api-access-99g7g\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.845890 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-utilities\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.946981 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-catalog-content\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.947026 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99g7g\" (UniqueName: \"kubernetes.io/projected/b4c13351-b8fa-4224-a09b-942200d398b1-kube-api-access-99g7g\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.947052 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-utilities\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.947531 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-utilities\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.947607 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-catalog-content\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:16 crc kubenswrapper[4778]: I0312 13:24:16.965068 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99g7g\" (UniqueName: \"kubernetes.io/projected/b4c13351-b8fa-4224-a09b-942200d398b1-kube-api-access-99g7g\") pod \"redhat-operators-tpgqw\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:17 crc kubenswrapper[4778]: I0312 13:24:17.072150 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:17 crc kubenswrapper[4778]: I0312 13:24:17.473882 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tpgqw"] Mar 12 13:24:17 crc kubenswrapper[4778]: W0312 13:24:17.481321 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4c13351_b8fa_4224_a09b_942200d398b1.slice/crio-18bee62b19e956e2c3d75695d30248e10ca9db095c911e5857cfc9283eb0b145 WatchSource:0}: Error finding container 18bee62b19e956e2c3d75695d30248e10ca9db095c911e5857cfc9283eb0b145: Status 404 returned error can't find the container with id 18bee62b19e956e2c3d75695d30248e10ca9db095c911e5857cfc9283eb0b145 Mar 12 13:24:18 crc kubenswrapper[4778]: I0312 13:24:18.235134 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4c13351-b8fa-4224-a09b-942200d398b1" containerID="24fd7a17a52a8b61e1c1382d3f50546a92de510fde86d6489d611167d57ee7a2" exitCode=0 Mar 12 13:24:18 crc kubenswrapper[4778]: I0312 13:24:18.235218 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerDied","Data":"24fd7a17a52a8b61e1c1382d3f50546a92de510fde86d6489d611167d57ee7a2"} Mar 12 13:24:18 crc kubenswrapper[4778]: I0312 13:24:18.235630 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerStarted","Data":"18bee62b19e956e2c3d75695d30248e10ca9db095c911e5857cfc9283eb0b145"} Mar 12 13:24:18 crc kubenswrapper[4778]: I0312 13:24:18.237774 4778 generic.go:334] "Generic (PLEG): container finished" podID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerID="bcf5a20a1141fa74f4f4862777ae90c4f7c67228670a44fa61079af3b9275916" exitCode=0 Mar 12 13:24:18 crc kubenswrapper[4778]: I0312 13:24:18.237851 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" event={"ID":"cb93062b-8387-4eb4-8662-ecaf93146d85","Type":"ContainerDied","Data":"bcf5a20a1141fa74f4f4862777ae90c4f7c67228670a44fa61079af3b9275916"} Mar 12 13:24:19 crc kubenswrapper[4778]: I0312 13:24:19.244569 4778 generic.go:334] "Generic (PLEG): container finished" podID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerID="fa7bdea962d6a0545c6439de4df98f392b5c38e2f8c4021c7527609f3773d2fc" exitCode=0 Mar 12 13:24:19 crc kubenswrapper[4778]: I0312 13:24:19.244655 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" event={"ID":"cb93062b-8387-4eb4-8662-ecaf93146d85","Type":"ContainerDied","Data":"fa7bdea962d6a0545c6439de4df98f392b5c38e2f8c4021c7527609f3773d2fc"} Mar 12 13:24:19 crc kubenswrapper[4778]: I0312 13:24:19.246429 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerStarted","Data":"43f23e51ea1a4786666d9356631a20c154ceff599adf74c4b7fdae769a58165f"} Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.018757 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.092866 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-bundle\") pod \"cb93062b-8387-4eb4-8662-ecaf93146d85\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.093011 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm5dc\" (UniqueName: \"kubernetes.io/projected/cb93062b-8387-4eb4-8662-ecaf93146d85-kube-api-access-gm5dc\") pod \"cb93062b-8387-4eb4-8662-ecaf93146d85\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.093058 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-util\") pod \"cb93062b-8387-4eb4-8662-ecaf93146d85\" (UID: \"cb93062b-8387-4eb4-8662-ecaf93146d85\") " Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.097303 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-bundle" (OuterVolumeSpecName: "bundle") pod "cb93062b-8387-4eb4-8662-ecaf93146d85" (UID: "cb93062b-8387-4eb4-8662-ecaf93146d85"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.104810 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-util" (OuterVolumeSpecName: "util") pod "cb93062b-8387-4eb4-8662-ecaf93146d85" (UID: "cb93062b-8387-4eb4-8662-ecaf93146d85"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.108201 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb93062b-8387-4eb4-8662-ecaf93146d85-kube-api-access-gm5dc" (OuterVolumeSpecName: "kube-api-access-gm5dc") pod "cb93062b-8387-4eb4-8662-ecaf93146d85" (UID: "cb93062b-8387-4eb4-8662-ecaf93146d85"). InnerVolumeSpecName "kube-api-access-gm5dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.195146 4778 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-util\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.195289 4778 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb93062b-8387-4eb4-8662-ecaf93146d85-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.195313 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm5dc\" (UniqueName: \"kubernetes.io/projected/cb93062b-8387-4eb4-8662-ecaf93146d85-kube-api-access-gm5dc\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.395970 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4c13351-b8fa-4224-a09b-942200d398b1" containerID="43f23e51ea1a4786666d9356631a20c154ceff599adf74c4b7fdae769a58165f" exitCode=0 Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.396031 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerDied","Data":"43f23e51ea1a4786666d9356631a20c154ceff599adf74c4b7fdae769a58165f"} Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.401476 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" event={"ID":"cb93062b-8387-4eb4-8662-ecaf93146d85","Type":"ContainerDied","Data":"3f052fe9978b090dcb12eafc251984a2e5ac0055088087a2227d82364c78a56d"} Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.401819 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f052fe9978b090dcb12eafc251984a2e5ac0055088087a2227d82364c78a56d" Mar 12 13:24:21 crc kubenswrapper[4778]: I0312 13:24:21.401619 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd" Mar 12 13:24:23 crc kubenswrapper[4778]: I0312 13:24:23.415017 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerStarted","Data":"4afc7793b8453c98e2325cef91c55591417e0eb2c857e8d9e7956e6b80421763"} Mar 12 13:24:23 crc kubenswrapper[4778]: I0312 13:24:23.433670 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tpgqw" podStartSLOduration=3.618079144 podStartE2EDuration="7.433644474s" podCreationTimestamp="2026-03-12 13:24:16 +0000 UTC" firstStartedPulling="2026-03-12 13:24:18.237156929 +0000 UTC m=+876.685852325" lastFinishedPulling="2026-03-12 13:24:22.052722259 +0000 UTC m=+880.501417655" observedRunningTime="2026-03-12 13:24:23.43070257 +0000 UTC m=+881.879397966" watchObservedRunningTime="2026-03-12 13:24:23.433644474 +0000 UTC m=+881.882339870" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.012503 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6"] Mar 12 13:24:25 crc kubenswrapper[4778]: E0312 13:24:25.012998 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="pull" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.013014 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="pull" Mar 12 13:24:25 crc kubenswrapper[4778]: E0312 13:24:25.013028 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="extract" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.013034 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="extract" Mar 12 13:24:25 crc kubenswrapper[4778]: E0312 13:24:25.013044 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="util" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.013051 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="util" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.013186 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb93062b-8387-4eb4-8662-ecaf93146d85" containerName="extract" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.013654 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.016418 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mzbl4" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.017611 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.017872 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.026068 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6"] Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.123731 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4rwl\" (UniqueName: \"kubernetes.io/projected/fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3-kube-api-access-q4rwl\") pod \"nmstate-operator-796d4cfff4-hxzd6\" (UID: \"fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.224770 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4rwl\" (UniqueName: \"kubernetes.io/projected/fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3-kube-api-access-q4rwl\") pod \"nmstate-operator-796d4cfff4-hxzd6\" (UID: \"fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.245299 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4rwl\" (UniqueName: \"kubernetes.io/projected/fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3-kube-api-access-q4rwl\") pod \"nmstate-operator-796d4cfff4-hxzd6\" (UID: \"fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.333699 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" Mar 12 13:24:25 crc kubenswrapper[4778]: I0312 13:24:25.935964 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6"] Mar 12 13:24:26 crc kubenswrapper[4778]: I0312 13:24:26.431520 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" event={"ID":"fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3","Type":"ContainerStarted","Data":"de7b463aeceaf4376c9af28c526743bec9c020efd46470d4686f4a4e81971bd2"} Mar 12 13:24:27 crc kubenswrapper[4778]: I0312 13:24:27.072686 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:27 crc kubenswrapper[4778]: I0312 13:24:27.072738 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:28 crc kubenswrapper[4778]: I0312 13:24:28.108847 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tpgqw" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="registry-server" probeResult="failure" output=< Mar 12 13:24:28 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:24:28 crc kubenswrapper[4778]: > Mar 12 13:24:32 crc kubenswrapper[4778]: I0312 13:24:32.466191 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" event={"ID":"fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3","Type":"ContainerStarted","Data":"ebf1c0e7ad87a1fe1a79167522049e1d248c464828ae36b67d71839bfdf46854"} Mar 12 13:24:32 crc kubenswrapper[4778]: I0312 13:24:32.484249 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-hxzd6" podStartSLOduration=3.105399124 podStartE2EDuration="8.484216447s" podCreationTimestamp="2026-03-12 13:24:24 +0000 UTC" firstStartedPulling="2026-03-12 13:24:25.954441143 +0000 UTC m=+884.403136539" lastFinishedPulling="2026-03-12 13:24:31.333258466 +0000 UTC m=+889.781953862" observedRunningTime="2026-03-12 13:24:32.480773129 +0000 UTC m=+890.929468535" watchObservedRunningTime="2026-03-12 13:24:32.484216447 +0000 UTC m=+890.932911843" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.676123 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.677339 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.684099 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-hj2n7" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.690572 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.709705 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-94rbc"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.710655 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.712800 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.732391 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef796a94-b10d-4d18-ae88-f64bc3a6b87d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-94rbc\" (UID: \"ef796a94-b10d-4d18-ae88-f64bc3a6b87d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.732664 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92frx\" (UniqueName: \"kubernetes.io/projected/ef796a94-b10d-4d18-ae88-f64bc3a6b87d-kube-api-access-92frx\") pod \"nmstate-webhook-5f558f5558-94rbc\" (UID: \"ef796a94-b10d-4d18-ae88-f64bc3a6b87d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.732821 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6mcn\" (UniqueName: \"kubernetes.io/projected/7855d7b1-c7cf-4b63-9313-051a391fcf43-kube-api-access-c6mcn\") pod \"nmstate-metrics-9b8c8685d-b2s5h\" (UID: \"7855d7b1-c7cf-4b63-9313-051a391fcf43\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.771484 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-94rbc"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.782555 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-rbsjl"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.783757 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.833544 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klgt6\" (UniqueName: \"kubernetes.io/projected/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-kube-api-access-klgt6\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.833920 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef796a94-b10d-4d18-ae88-f64bc3a6b87d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-94rbc\" (UID: \"ef796a94-b10d-4d18-ae88-f64bc3a6b87d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.833946 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-dbus-socket\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.833975 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92frx\" (UniqueName: \"kubernetes.io/projected/ef796a94-b10d-4d18-ae88-f64bc3a6b87d-kube-api-access-92frx\") pod \"nmstate-webhook-5f558f5558-94rbc\" (UID: \"ef796a94-b10d-4d18-ae88-f64bc3a6b87d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.834005 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-nmstate-lock\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.835079 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6mcn\" (UniqueName: \"kubernetes.io/projected/7855d7b1-c7cf-4b63-9313-051a391fcf43-kube-api-access-c6mcn\") pod \"nmstate-metrics-9b8c8685d-b2s5h\" (UID: \"7855d7b1-c7cf-4b63-9313-051a391fcf43\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.835343 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-ovs-socket\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.843948 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef796a94-b10d-4d18-ae88-f64bc3a6b87d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-94rbc\" (UID: \"ef796a94-b10d-4d18-ae88-f64bc3a6b87d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.851332 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.852159 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.859986 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.860935 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-s925z" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.861160 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.867338 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6mcn\" (UniqueName: \"kubernetes.io/projected/7855d7b1-c7cf-4b63-9313-051a391fcf43-kube-api-access-c6mcn\") pod \"nmstate-metrics-9b8c8685d-b2s5h\" (UID: \"7855d7b1-c7cf-4b63-9313-051a391fcf43\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.867594 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4"] Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.883009 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92frx\" (UniqueName: \"kubernetes.io/projected/ef796a94-b10d-4d18-ae88-f64bc3a6b87d-kube-api-access-92frx\") pod \"nmstate-webhook-5f558f5558-94rbc\" (UID: \"ef796a94-b10d-4d18-ae88-f64bc3a6b87d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.936088 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8whd5\" (UniqueName: \"kubernetes.io/projected/af2d568b-9719-4da9-b0e8-e28d314ed860-kube-api-access-8whd5\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.936454 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klgt6\" (UniqueName: \"kubernetes.io/projected/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-kube-api-access-klgt6\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.936944 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-dbus-socket\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.937497 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/af2d568b-9719-4da9-b0e8-e28d314ed860-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.937661 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-nmstate-lock\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.937831 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/af2d568b-9719-4da9-b0e8-e28d314ed860-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.937967 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-ovs-socket\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.937382 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-dbus-socket\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.937772 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-nmstate-lock\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.938089 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-ovs-socket\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.954581 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klgt6\" (UniqueName: \"kubernetes.io/projected/d8309ffe-a26c-44a8-84e2-7b7ec10982a8-kube-api-access-klgt6\") pod \"nmstate-handler-rbsjl\" (UID: \"d8309ffe-a26c-44a8-84e2-7b7ec10982a8\") " pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:34 crc kubenswrapper[4778]: I0312 13:24:34.997168 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.023174 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.038776 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/af2d568b-9719-4da9-b0e8-e28d314ed860-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.038858 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8whd5\" (UniqueName: \"kubernetes.io/projected/af2d568b-9719-4da9-b0e8-e28d314ed860-kube-api-access-8whd5\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.038913 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/af2d568b-9719-4da9-b0e8-e28d314ed860-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: E0312 13:24:35.039058 4778 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 12 13:24:35 crc kubenswrapper[4778]: E0312 13:24:35.039130 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af2d568b-9719-4da9-b0e8-e28d314ed860-plugin-serving-cert podName:af2d568b-9719-4da9-b0e8-e28d314ed860 nodeName:}" failed. No retries permitted until 2026-03-12 13:24:35.539098969 +0000 UTC m=+893.987794365 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/af2d568b-9719-4da9-b0e8-e28d314ed860-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-jbxx4" (UID: "af2d568b-9719-4da9-b0e8-e28d314ed860") : secret "plugin-serving-cert" not found Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.040160 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/af2d568b-9719-4da9-b0e8-e28d314ed860-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.060763 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8whd5\" (UniqueName: \"kubernetes.io/projected/af2d568b-9719-4da9-b0e8-e28d314ed860-kube-api-access-8whd5\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.107985 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.129279 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-958f4d6df-8h8cb"] Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.130456 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.139789 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-958f4d6df-8h8cb"] Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.140807 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-oauth-serving-cert\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: W0312 13:24:35.140930 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8309ffe_a26c_44a8_84e2_7b7ec10982a8.slice/crio-81b278cc5e7a181ca087fed20e2d8d360596a3f92009d968b804af7d71137571 WatchSource:0}: Error finding container 81b278cc5e7a181ca087fed20e2d8d360596a3f92009d968b804af7d71137571: Status 404 returned error can't find the container with id 81b278cc5e7a181ca087fed20e2d8d360596a3f92009d968b804af7d71137571 Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.141574 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwbvd\" (UniqueName: \"kubernetes.io/projected/21091ec0-7369-4f02-949c-ca61ed2efad3-kube-api-access-hwbvd\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.141875 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-trusted-ca-bundle\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.142043 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/21091ec0-7369-4f02-949c-ca61ed2efad3-console-oauth-config\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.142211 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/21091ec0-7369-4f02-949c-ca61ed2efad3-console-serving-cert\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.142328 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-console-config\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.142557 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-service-ca\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.244091 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-oauth-serving-cert\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.244477 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwbvd\" (UniqueName: \"kubernetes.io/projected/21091ec0-7369-4f02-949c-ca61ed2efad3-kube-api-access-hwbvd\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.244508 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-trusted-ca-bundle\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.244865 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/21091ec0-7369-4f02-949c-ca61ed2efad3-console-oauth-config\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.245516 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-oauth-serving-cert\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.245743 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/21091ec0-7369-4f02-949c-ca61ed2efad3-console-serving-cert\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.245787 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-console-config\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.245827 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-service-ca\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.247815 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-trusted-ca-bundle\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.247888 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-console-config\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.248737 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/21091ec0-7369-4f02-949c-ca61ed2efad3-service-ca\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.253102 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/21091ec0-7369-4f02-949c-ca61ed2efad3-console-oauth-config\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.255213 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/21091ec0-7369-4f02-949c-ca61ed2efad3-console-serving-cert\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.266045 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwbvd\" (UniqueName: \"kubernetes.io/projected/21091ec0-7369-4f02-949c-ca61ed2efad3-kube-api-access-hwbvd\") pod \"console-958f4d6df-8h8cb\" (UID: \"21091ec0-7369-4f02-949c-ca61ed2efad3\") " pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.577339 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.579158 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/af2d568b-9719-4da9-b0e8-e28d314ed860-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.582021 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/af2d568b-9719-4da9-b0e8-e28d314ed860-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jbxx4\" (UID: \"af2d568b-9719-4da9-b0e8-e28d314ed860\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.587714 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rbsjl" event={"ID":"d8309ffe-a26c-44a8-84e2-7b7ec10982a8","Type":"ContainerStarted","Data":"81b278cc5e7a181ca087fed20e2d8d360596a3f92009d968b804af7d71137571"} Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.801148 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h"] Mar 12 13:24:35 crc kubenswrapper[4778]: W0312 13:24:35.814354 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7855d7b1_c7cf_4b63_9313_051a391fcf43.slice/crio-d895f3bf40dc0ae5f48efaff396a677d277046ea34568bfb3a907c76dcd211da WatchSource:0}: Error finding container d895f3bf40dc0ae5f48efaff396a677d277046ea34568bfb3a907c76dcd211da: Status 404 returned error can't find the container with id d895f3bf40dc0ae5f48efaff396a677d277046ea34568bfb3a907c76dcd211da Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.816675 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.899053 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-94rbc"] Mar 12 13:24:35 crc kubenswrapper[4778]: I0312 13:24:35.969435 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-958f4d6df-8h8cb"] Mar 12 13:24:35 crc kubenswrapper[4778]: W0312 13:24:35.988200 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21091ec0_7369_4f02_949c_ca61ed2efad3.slice/crio-3e63c6e680e69252e84320ddf99ca63c93d07c8fd23ecc062ca7cc30ce720b4a WatchSource:0}: Error finding container 3e63c6e680e69252e84320ddf99ca63c93d07c8fd23ecc062ca7cc30ce720b4a: Status 404 returned error can't find the container with id 3e63c6e680e69252e84320ddf99ca63c93d07c8fd23ecc062ca7cc30ce720b4a Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.076229 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4"] Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.595434 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" event={"ID":"af2d568b-9719-4da9-b0e8-e28d314ed860","Type":"ContainerStarted","Data":"3930912cf9cf2867c2bf1870b08d2e416418de2845bd14af9b08e8f4e4e8ffda"} Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.597051 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-958f4d6df-8h8cb" event={"ID":"21091ec0-7369-4f02-949c-ca61ed2efad3","Type":"ContainerStarted","Data":"64d95a47cc1d737b6ce99e7266eadeebc2fd758b4619d4cd8bf19acb253bf593"} Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.597080 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-958f4d6df-8h8cb" event={"ID":"21091ec0-7369-4f02-949c-ca61ed2efad3","Type":"ContainerStarted","Data":"3e63c6e680e69252e84320ddf99ca63c93d07c8fd23ecc062ca7cc30ce720b4a"} Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.598193 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" event={"ID":"ef796a94-b10d-4d18-ae88-f64bc3a6b87d","Type":"ContainerStarted","Data":"6699f1ed9504b8bd4081021164004700a3296a23da446a20b0f679599f43ed90"} Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.600023 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" event={"ID":"7855d7b1-c7cf-4b63-9313-051a391fcf43","Type":"ContainerStarted","Data":"d895f3bf40dc0ae5f48efaff396a677d277046ea34568bfb3a907c76dcd211da"} Mar 12 13:24:36 crc kubenswrapper[4778]: I0312 13:24:36.620286 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-958f4d6df-8h8cb" podStartSLOduration=1.620259022 podStartE2EDuration="1.620259022s" podCreationTimestamp="2026-03-12 13:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:24:36.616202516 +0000 UTC m=+895.064897912" watchObservedRunningTime="2026-03-12 13:24:36.620259022 +0000 UTC m=+895.068954418" Mar 12 13:24:37 crc kubenswrapper[4778]: I0312 13:24:37.113985 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:37 crc kubenswrapper[4778]: I0312 13:24:37.162095 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:37 crc kubenswrapper[4778]: I0312 13:24:37.350992 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tpgqw"] Mar 12 13:24:38 crc kubenswrapper[4778]: I0312 13:24:38.653734 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tpgqw" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="registry-server" containerID="cri-o://4afc7793b8453c98e2325cef91c55591417e0eb2c857e8d9e7956e6b80421763" gracePeriod=2 Mar 12 13:24:39 crc kubenswrapper[4778]: I0312 13:24:39.708177 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4c13351-b8fa-4224-a09b-942200d398b1" containerID="4afc7793b8453c98e2325cef91c55591417e0eb2c857e8d9e7956e6b80421763" exitCode=0 Mar 12 13:24:39 crc kubenswrapper[4778]: I0312 13:24:39.708273 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerDied","Data":"4afc7793b8453c98e2325cef91c55591417e0eb2c857e8d9e7956e6b80421763"} Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.295611 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.466209 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-catalog-content\") pod \"b4c13351-b8fa-4224-a09b-942200d398b1\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.466842 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99g7g\" (UniqueName: \"kubernetes.io/projected/b4c13351-b8fa-4224-a09b-942200d398b1-kube-api-access-99g7g\") pod \"b4c13351-b8fa-4224-a09b-942200d398b1\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.466891 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-utilities\") pod \"b4c13351-b8fa-4224-a09b-942200d398b1\" (UID: \"b4c13351-b8fa-4224-a09b-942200d398b1\") " Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.467705 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-utilities" (OuterVolumeSpecName: "utilities") pod "b4c13351-b8fa-4224-a09b-942200d398b1" (UID: "b4c13351-b8fa-4224-a09b-942200d398b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.472792 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4c13351-b8fa-4224-a09b-942200d398b1-kube-api-access-99g7g" (OuterVolumeSpecName: "kube-api-access-99g7g") pod "b4c13351-b8fa-4224-a09b-942200d398b1" (UID: "b4c13351-b8fa-4224-a09b-942200d398b1"). InnerVolumeSpecName "kube-api-access-99g7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.568589 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.568618 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99g7g\" (UniqueName: \"kubernetes.io/projected/b4c13351-b8fa-4224-a09b-942200d398b1-kube-api-access-99g7g\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.596482 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4c13351-b8fa-4224-a09b-942200d398b1" (UID: "b4c13351-b8fa-4224-a09b-942200d398b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.669600 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c13351-b8fa-4224-a09b-942200d398b1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.724253 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" event={"ID":"7855d7b1-c7cf-4b63-9313-051a391fcf43","Type":"ContainerStarted","Data":"3d8aa893a82e4d0fc1feb25640d8f5da687769e944fb0bef634f4a574a727877"} Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.726768 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" event={"ID":"ef796a94-b10d-4d18-ae88-f64bc3a6b87d","Type":"ContainerStarted","Data":"dd669951dbca11310165ad2a5acc960c5b8032a1f80ebce35c75a79f6d31256d"} Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.726921 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.730163 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tpgqw" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.730169 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tpgqw" event={"ID":"b4c13351-b8fa-4224-a09b-942200d398b1","Type":"ContainerDied","Data":"18bee62b19e956e2c3d75695d30248e10ca9db095c911e5857cfc9283eb0b145"} Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.730249 4778 scope.go:117] "RemoveContainer" containerID="4afc7793b8453c98e2325cef91c55591417e0eb2c857e8d9e7956e6b80421763" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.731914 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rbsjl" event={"ID":"d8309ffe-a26c-44a8-84e2-7b7ec10982a8","Type":"ContainerStarted","Data":"29d434fae31803af5a7f922b85304b26cfa18b09b618b2f8ebeda48d234d1a6f"} Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.732085 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.734243 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" event={"ID":"af2d568b-9719-4da9-b0e8-e28d314ed860","Type":"ContainerStarted","Data":"be918cffe388f198c074ff0e1b9ef80800945389cd3920982e304a43c77394ec"} Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.759603 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" podStartSLOduration=2.325521461 podStartE2EDuration="7.759572917s" podCreationTimestamp="2026-03-12 13:24:34 +0000 UTC" firstStartedPulling="2026-03-12 13:24:35.920357791 +0000 UTC m=+894.369053187" lastFinishedPulling="2026-03-12 13:24:41.354409247 +0000 UTC m=+899.803104643" observedRunningTime="2026-03-12 13:24:41.753972928 +0000 UTC m=+900.202668334" watchObservedRunningTime="2026-03-12 13:24:41.759572917 +0000 UTC m=+900.208268313" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.763397 4778 scope.go:117] "RemoveContainer" containerID="43f23e51ea1a4786666d9356631a20c154ceff599adf74c4b7fdae769a58165f" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.780269 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-rbsjl" podStartSLOduration=1.574613844 podStartE2EDuration="7.780239276s" podCreationTimestamp="2026-03-12 13:24:34 +0000 UTC" firstStartedPulling="2026-03-12 13:24:35.150128174 +0000 UTC m=+893.598823580" lastFinishedPulling="2026-03-12 13:24:41.355753616 +0000 UTC m=+899.804449012" observedRunningTime="2026-03-12 13:24:41.777523889 +0000 UTC m=+900.226219295" watchObservedRunningTime="2026-03-12 13:24:41.780239276 +0000 UTC m=+900.228934672" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.801200 4778 scope.go:117] "RemoveContainer" containerID="24fd7a17a52a8b61e1c1382d3f50546a92de510fde86d6489d611167d57ee7a2" Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.806974 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tpgqw"] Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.838739 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tpgqw"] Mar 12 13:24:41 crc kubenswrapper[4778]: I0312 13:24:41.846284 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jbxx4" podStartSLOduration=2.588493697 podStartE2EDuration="7.846247038s" podCreationTimestamp="2026-03-12 13:24:34 +0000 UTC" firstStartedPulling="2026-03-12 13:24:36.097894682 +0000 UTC m=+894.546590078" lastFinishedPulling="2026-03-12 13:24:41.355648023 +0000 UTC m=+899.804343419" observedRunningTime="2026-03-12 13:24:41.820633448 +0000 UTC m=+900.269328874" watchObservedRunningTime="2026-03-12 13:24:41.846247038 +0000 UTC m=+900.294942434" Mar 12 13:24:42 crc kubenswrapper[4778]: I0312 13:24:42.269761 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" path="/var/lib/kubelet/pods/b4c13351-b8fa-4224-a09b-942200d398b1/volumes" Mar 12 13:24:45 crc kubenswrapper[4778]: I0312 13:24:45.578056 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:45 crc kubenswrapper[4778]: I0312 13:24:45.578460 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:45 crc kubenswrapper[4778]: I0312 13:24:45.583703 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:45 crc kubenswrapper[4778]: I0312 13:24:45.767417 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-958f4d6df-8h8cb" Mar 12 13:24:45 crc kubenswrapper[4778]: I0312 13:24:45.821922 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xwwxp"] Mar 12 13:24:46 crc kubenswrapper[4778]: I0312 13:24:46.838678 4778 scope.go:117] "RemoveContainer" containerID="42a7fef965fea72fd4ae8fcc7e99e6b821d3626af8cb88a527c7193c956003a6" Mar 12 13:24:47 crc kubenswrapper[4778]: I0312 13:24:47.778068 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" event={"ID":"7855d7b1-c7cf-4b63-9313-051a391fcf43","Type":"ContainerStarted","Data":"2bad13e2195c34f7af1b387fa86c2c44b3f3c0d160ed233dff08d76f5e7daf6c"} Mar 12 13:24:50 crc kubenswrapper[4778]: I0312 13:24:50.135942 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-rbsjl" Mar 12 13:24:50 crc kubenswrapper[4778]: I0312 13:24:50.157028 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-b2s5h" podStartSLOduration=4.752099645 podStartE2EDuration="16.157000711s" podCreationTimestamp="2026-03-12 13:24:34 +0000 UTC" firstStartedPulling="2026-03-12 13:24:35.817503649 +0000 UTC m=+894.266199045" lastFinishedPulling="2026-03-12 13:24:47.222404715 +0000 UTC m=+905.671100111" observedRunningTime="2026-03-12 13:24:47.802081259 +0000 UTC m=+906.250776675" watchObservedRunningTime="2026-03-12 13:24:50.157000711 +0000 UTC m=+908.605696107" Mar 12 13:24:55 crc kubenswrapper[4778]: I0312 13:24:55.032741 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-94rbc" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.744692 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6"] Mar 12 13:25:07 crc kubenswrapper[4778]: E0312 13:25:07.745457 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="extract-content" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.745472 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="extract-content" Mar 12 13:25:07 crc kubenswrapper[4778]: E0312 13:25:07.745486 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="extract-utilities" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.745493 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="extract-utilities" Mar 12 13:25:07 crc kubenswrapper[4778]: E0312 13:25:07.745512 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="registry-server" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.745521 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="registry-server" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.745635 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4c13351-b8fa-4224-a09b-942200d398b1" containerName="registry-server" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.746367 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.748584 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.765423 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6"] Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.851960 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.852021 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47j9d\" (UniqueName: \"kubernetes.io/projected/9090029d-2f37-457b-8425-3690da177434-kube-api-access-47j9d\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.852061 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.953647 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.953992 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.954091 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47j9d\" (UniqueName: \"kubernetes.io/projected/9090029d-2f37-457b-8425-3690da177434-kube-api-access-47j9d\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.954130 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.954546 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:07 crc kubenswrapper[4778]: I0312 13:25:07.976098 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47j9d\" (UniqueName: \"kubernetes.io/projected/9090029d-2f37-457b-8425-3690da177434-kube-api-access-47j9d\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:08 crc kubenswrapper[4778]: I0312 13:25:08.068458 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:08 crc kubenswrapper[4778]: I0312 13:25:08.457094 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6"] Mar 12 13:25:08 crc kubenswrapper[4778]: I0312 13:25:08.900404 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" event={"ID":"9090029d-2f37-457b-8425-3690da177434","Type":"ContainerStarted","Data":"a21914d65f0222503b56911bb4a1dc512ff6b6f36daa27eac18a63f864ad2b02"} Mar 12 13:25:08 crc kubenswrapper[4778]: I0312 13:25:08.900455 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" event={"ID":"9090029d-2f37-457b-8425-3690da177434","Type":"ContainerStarted","Data":"b9aa60e4a4d8c2467aa722d6a95be03da066e99dbe5fc5eb08f4ee10964bfad1"} Mar 12 13:25:09 crc kubenswrapper[4778]: I0312 13:25:09.910232 4778 generic.go:334] "Generic (PLEG): container finished" podID="9090029d-2f37-457b-8425-3690da177434" containerID="a21914d65f0222503b56911bb4a1dc512ff6b6f36daa27eac18a63f864ad2b02" exitCode=0 Mar 12 13:25:09 crc kubenswrapper[4778]: I0312 13:25:09.910294 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" event={"ID":"9090029d-2f37-457b-8425-3690da177434","Type":"ContainerDied","Data":"a21914d65f0222503b56911bb4a1dc512ff6b6f36daa27eac18a63f864ad2b02"} Mar 12 13:25:10 crc kubenswrapper[4778]: I0312 13:25:10.863279 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-xwwxp" podUID="c825022c-79bc-44ae-bc64-ee9614aafe25" containerName="console" containerID="cri-o://4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c" gracePeriod=15 Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.226732 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xwwxp_c825022c-79bc-44ae-bc64-ee9614aafe25/console/0.log" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.226805 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.394842 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-serving-cert\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.395291 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-oauth-config\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.395326 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-oauth-serving-cert\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.395375 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqbjv\" (UniqueName: \"kubernetes.io/projected/c825022c-79bc-44ae-bc64-ee9614aafe25-kube-api-access-rqbjv\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.395433 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-service-ca\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.395478 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-trusted-ca-bundle\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.395514 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-console-config\") pod \"c825022c-79bc-44ae-bc64-ee9614aafe25\" (UID: \"c825022c-79bc-44ae-bc64-ee9614aafe25\") " Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.396529 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-console-config" (OuterVolumeSpecName: "console-config") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.396685 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.396715 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-service-ca" (OuterVolumeSpecName: "service-ca") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.397056 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.401085 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.401743 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.401805 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c825022c-79bc-44ae-bc64-ee9614aafe25-kube-api-access-rqbjv" (OuterVolumeSpecName: "kube-api-access-rqbjv") pod "c825022c-79bc-44ae-bc64-ee9614aafe25" (UID: "c825022c-79bc-44ae-bc64-ee9614aafe25"). InnerVolumeSpecName "kube-api-access-rqbjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497365 4778 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-console-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497441 4778 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497467 4778 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c825022c-79bc-44ae-bc64-ee9614aafe25-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497493 4778 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497516 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqbjv\" (UniqueName: \"kubernetes.io/projected/c825022c-79bc-44ae-bc64-ee9614aafe25-kube-api-access-rqbjv\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497542 4778 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.497564 4778 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c825022c-79bc-44ae-bc64-ee9614aafe25-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.920683 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xwwxp_c825022c-79bc-44ae-bc64-ee9614aafe25/console/0.log" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.920724 4778 generic.go:334] "Generic (PLEG): container finished" podID="c825022c-79bc-44ae-bc64-ee9614aafe25" containerID="4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c" exitCode=2 Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.920748 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwwxp" event={"ID":"c825022c-79bc-44ae-bc64-ee9614aafe25","Type":"ContainerDied","Data":"4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c"} Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.920774 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwwxp" event={"ID":"c825022c-79bc-44ae-bc64-ee9614aafe25","Type":"ContainerDied","Data":"6f20116905733a7dbe8802503613a6b31a51c117f53f02f55e4cace656d26f20"} Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.920791 4778 scope.go:117] "RemoveContainer" containerID="4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.920792 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwwxp" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.937327 4778 scope.go:117] "RemoveContainer" containerID="4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c" Mar 12 13:25:11 crc kubenswrapper[4778]: E0312 13:25:11.937793 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c\": container with ID starting with 4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c not found: ID does not exist" containerID="4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.937828 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c"} err="failed to get container status \"4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c\": rpc error: code = NotFound desc = could not find container \"4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c\": container with ID starting with 4f4a64269de7f325ca6cad0c8f6bdffa97bc955d4a92c8f27548dcfdbd421f4c not found: ID does not exist" Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.953715 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xwwxp"] Mar 12 13:25:11 crc kubenswrapper[4778]: I0312 13:25:11.958118 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-xwwxp"] Mar 12 13:25:12 crc kubenswrapper[4778]: I0312 13:25:12.262680 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c825022c-79bc-44ae-bc64-ee9614aafe25" path="/var/lib/kubelet/pods/c825022c-79bc-44ae-bc64-ee9614aafe25/volumes" Mar 12 13:25:12 crc kubenswrapper[4778]: I0312 13:25:12.931213 4778 generic.go:334] "Generic (PLEG): container finished" podID="9090029d-2f37-457b-8425-3690da177434" containerID="815dcddc546761ce579fc83410d6b0089762cd2da7d4030454ea7174a833b745" exitCode=0 Mar 12 13:25:12 crc kubenswrapper[4778]: I0312 13:25:12.931292 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" event={"ID":"9090029d-2f37-457b-8425-3690da177434","Type":"ContainerDied","Data":"815dcddc546761ce579fc83410d6b0089762cd2da7d4030454ea7174a833b745"} Mar 12 13:25:13 crc kubenswrapper[4778]: I0312 13:25:13.943365 4778 generic.go:334] "Generic (PLEG): container finished" podID="9090029d-2f37-457b-8425-3690da177434" containerID="e9a2263250dc38f09156c0b8d5a3489b7c876f2a032e5671b5417dd730d61037" exitCode=0 Mar 12 13:25:13 crc kubenswrapper[4778]: I0312 13:25:13.943441 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" event={"ID":"9090029d-2f37-457b-8425-3690da177434","Type":"ContainerDied","Data":"e9a2263250dc38f09156c0b8d5a3489b7c876f2a032e5671b5417dd730d61037"} Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.140197 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.239450 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47j9d\" (UniqueName: \"kubernetes.io/projected/9090029d-2f37-457b-8425-3690da177434-kube-api-access-47j9d\") pod \"9090029d-2f37-457b-8425-3690da177434\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.239559 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-bundle\") pod \"9090029d-2f37-457b-8425-3690da177434\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.239607 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-util\") pod \"9090029d-2f37-457b-8425-3690da177434\" (UID: \"9090029d-2f37-457b-8425-3690da177434\") " Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.240809 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-bundle" (OuterVolumeSpecName: "bundle") pod "9090029d-2f37-457b-8425-3690da177434" (UID: "9090029d-2f37-457b-8425-3690da177434"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.244441 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9090029d-2f37-457b-8425-3690da177434-kube-api-access-47j9d" (OuterVolumeSpecName: "kube-api-access-47j9d") pod "9090029d-2f37-457b-8425-3690da177434" (UID: "9090029d-2f37-457b-8425-3690da177434"). InnerVolumeSpecName "kube-api-access-47j9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.250550 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-util" (OuterVolumeSpecName: "util") pod "9090029d-2f37-457b-8425-3690da177434" (UID: "9090029d-2f37-457b-8425-3690da177434"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.341427 4778 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-util\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.341464 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47j9d\" (UniqueName: \"kubernetes.io/projected/9090029d-2f37-457b-8425-3690da177434-kube-api-access-47j9d\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.341478 4778 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9090029d-2f37-457b-8425-3690da177434-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.955578 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" event={"ID":"9090029d-2f37-457b-8425-3690da177434","Type":"ContainerDied","Data":"b9aa60e4a4d8c2467aa722d6a95be03da066e99dbe5fc5eb08f4ee10964bfad1"} Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.955911 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9aa60e4a4d8c2467aa722d6a95be03da066e99dbe5fc5eb08f4ee10964bfad1" Mar 12 13:25:15 crc kubenswrapper[4778]: I0312 13:25:15.955978 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.715170 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx"] Mar 12 13:25:26 crc kubenswrapper[4778]: E0312 13:25:26.716394 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c825022c-79bc-44ae-bc64-ee9614aafe25" containerName="console" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.716413 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c825022c-79bc-44ae-bc64-ee9614aafe25" containerName="console" Mar 12 13:25:26 crc kubenswrapper[4778]: E0312 13:25:26.716439 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="pull" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.716447 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="pull" Mar 12 13:25:26 crc kubenswrapper[4778]: E0312 13:25:26.716474 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="util" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.716482 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="util" Mar 12 13:25:26 crc kubenswrapper[4778]: E0312 13:25:26.716500 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="extract" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.716507 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="extract" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.716854 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c825022c-79bc-44ae-bc64-ee9614aafe25" containerName="console" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.716880 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9090029d-2f37-457b-8425-3690da177434" containerName="extract" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.717650 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.728840 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.732025 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.738634 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.738918 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.739246 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rk9jh" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.749816 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx"] Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.785873 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt2z4\" (UniqueName: \"kubernetes.io/projected/a5a6d344-0a75-422d-acd9-fe8887b03110-kube-api-access-kt2z4\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.786064 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5a6d344-0a75-422d-acd9-fe8887b03110-webhook-cert\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.786169 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5a6d344-0a75-422d-acd9-fe8887b03110-apiservice-cert\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.888213 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5a6d344-0a75-422d-acd9-fe8887b03110-apiservice-cert\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.889789 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt2z4\" (UniqueName: \"kubernetes.io/projected/a5a6d344-0a75-422d-acd9-fe8887b03110-kube-api-access-kt2z4\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.889915 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5a6d344-0a75-422d-acd9-fe8887b03110-webhook-cert\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.896114 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5a6d344-0a75-422d-acd9-fe8887b03110-webhook-cert\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.896151 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5a6d344-0a75-422d-acd9-fe8887b03110-apiservice-cert\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.907475 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt2z4\" (UniqueName: \"kubernetes.io/projected/a5a6d344-0a75-422d-acd9-fe8887b03110-kube-api-access-kt2z4\") pod \"metallb-operator-controller-manager-54d5c4b6c7-gh4lx\" (UID: \"a5a6d344-0a75-422d-acd9-fe8887b03110\") " pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.973511 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq"] Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.974811 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.977278 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.978056 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-k2kwx" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.978086 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.990672 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-apiservice-cert\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.990735 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-webhook-cert\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.990793 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwnxj\" (UniqueName: \"kubernetes.io/projected/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-kube-api-access-rwnxj\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:26 crc kubenswrapper[4778]: I0312 13:25:26.996825 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq"] Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.049632 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.091509 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-apiservice-cert\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.091590 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-webhook-cert\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.091637 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwnxj\" (UniqueName: \"kubernetes.io/projected/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-kube-api-access-rwnxj\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.096856 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-webhook-cert\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.097436 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-apiservice-cert\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.111089 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwnxj\" (UniqueName: \"kubernetes.io/projected/6ac207b6-1710-47af-8fe9-b0c3adbce0ab-kube-api-access-rwnxj\") pod \"metallb-operator-webhook-server-68f5db54d6-zstmq\" (UID: \"6ac207b6-1710-47af-8fe9-b0c3adbce0ab\") " pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.291460 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.686992 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx"] Mar 12 13:25:27 crc kubenswrapper[4778]: I0312 13:25:27.913484 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq"] Mar 12 13:25:27 crc kubenswrapper[4778]: W0312 13:25:27.924756 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ac207b6_1710_47af_8fe9_b0c3adbce0ab.slice/crio-6b6435c5a162a19fbf3fd787d2bd4b066257f5690be74c8c365417850ccb0c59 WatchSource:0}: Error finding container 6b6435c5a162a19fbf3fd787d2bd4b066257f5690be74c8c365417850ccb0c59: Status 404 returned error can't find the container with id 6b6435c5a162a19fbf3fd787d2bd4b066257f5690be74c8c365417850ccb0c59 Mar 12 13:25:28 crc kubenswrapper[4778]: I0312 13:25:28.054978 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" event={"ID":"a5a6d344-0a75-422d-acd9-fe8887b03110","Type":"ContainerStarted","Data":"736b3dbbffc16edb5fb0ca90cdffd191d6903c523af75123ba414476c8c2973b"} Mar 12 13:25:28 crc kubenswrapper[4778]: I0312 13:25:28.056415 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" event={"ID":"6ac207b6-1710-47af-8fe9-b0c3adbce0ab","Type":"ContainerStarted","Data":"6b6435c5a162a19fbf3fd787d2bd4b066257f5690be74c8c365417850ccb0c59"} Mar 12 13:25:28 crc kubenswrapper[4778]: I0312 13:25:28.558250 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:25:28 crc kubenswrapper[4778]: I0312 13:25:28.558649 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:25:38 crc kubenswrapper[4778]: I0312 13:25:38.237231 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" event={"ID":"6ac207b6-1710-47af-8fe9-b0c3adbce0ab","Type":"ContainerStarted","Data":"e3ec90eb649c9b7215dd5d53c7c610b2c43b328997544f79890d00ff5263b8fd"} Mar 12 13:25:38 crc kubenswrapper[4778]: I0312 13:25:38.237755 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:38 crc kubenswrapper[4778]: I0312 13:25:38.238800 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" event={"ID":"a5a6d344-0a75-422d-acd9-fe8887b03110","Type":"ContainerStarted","Data":"f8433cd677712c90f8cdb14e6420e1a6dea126ed1733e27ae61ed8dd45da56b4"} Mar 12 13:25:38 crc kubenswrapper[4778]: I0312 13:25:38.238971 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:25:38 crc kubenswrapper[4778]: I0312 13:25:38.264110 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" podStartSLOduration=2.9697980189999997 podStartE2EDuration="12.26408576s" podCreationTimestamp="2026-03-12 13:25:26 +0000 UTC" firstStartedPulling="2026-03-12 13:25:27.927514978 +0000 UTC m=+946.376210374" lastFinishedPulling="2026-03-12 13:25:37.221802719 +0000 UTC m=+955.670498115" observedRunningTime="2026-03-12 13:25:38.259649424 +0000 UTC m=+956.708344820" watchObservedRunningTime="2026-03-12 13:25:38.26408576 +0000 UTC m=+956.712781156" Mar 12 13:25:38 crc kubenswrapper[4778]: I0312 13:25:38.285705 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" podStartSLOduration=2.782420308 podStartE2EDuration="12.285691006s" podCreationTimestamp="2026-03-12 13:25:26 +0000 UTC" firstStartedPulling="2026-03-12 13:25:27.698334325 +0000 UTC m=+946.147029721" lastFinishedPulling="2026-03-12 13:25:37.201605023 +0000 UTC m=+955.650300419" observedRunningTime="2026-03-12 13:25:38.283636278 +0000 UTC m=+956.732331674" watchObservedRunningTime="2026-03-12 13:25:38.285691006 +0000 UTC m=+956.734386392" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.278935 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pltxt"] Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.280578 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.288540 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pltxt"] Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.414124 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-utilities\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.414315 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-catalog-content\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.414484 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b8rt\" (UniqueName: \"kubernetes.io/projected/3418515c-3077-4237-9aa9-596ed9d3c137-kube-api-access-2b8rt\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.515331 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-catalog-content\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.515423 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b8rt\" (UniqueName: \"kubernetes.io/projected/3418515c-3077-4237-9aa9-596ed9d3c137-kube-api-access-2b8rt\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.515473 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-utilities\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.515875 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-catalog-content\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.515901 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-utilities\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.537618 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b8rt\" (UniqueName: \"kubernetes.io/projected/3418515c-3077-4237-9aa9-596ed9d3c137-kube-api-access-2b8rt\") pod \"redhat-marketplace-pltxt\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.595022 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:41 crc kubenswrapper[4778]: I0312 13:25:41.919330 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pltxt"] Mar 12 13:25:41 crc kubenswrapper[4778]: W0312 13:25:41.925760 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3418515c_3077_4237_9aa9_596ed9d3c137.slice/crio-b37770ad61325c6792ceeea5e8eb5ca55edd1557d07b2fc176c76ded639b7940 WatchSource:0}: Error finding container b37770ad61325c6792ceeea5e8eb5ca55edd1557d07b2fc176c76ded639b7940: Status 404 returned error can't find the container with id b37770ad61325c6792ceeea5e8eb5ca55edd1557d07b2fc176c76ded639b7940 Mar 12 13:25:42 crc kubenswrapper[4778]: I0312 13:25:42.260632 4778 generic.go:334] "Generic (PLEG): container finished" podID="3418515c-3077-4237-9aa9-596ed9d3c137" containerID="6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864" exitCode=0 Mar 12 13:25:42 crc kubenswrapper[4778]: I0312 13:25:42.269626 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerDied","Data":"6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864"} Mar 12 13:25:42 crc kubenswrapper[4778]: I0312 13:25:42.269700 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerStarted","Data":"b37770ad61325c6792ceeea5e8eb5ca55edd1557d07b2fc176c76ded639b7940"} Mar 12 13:25:43 crc kubenswrapper[4778]: I0312 13:25:43.267697 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerStarted","Data":"33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a"} Mar 12 13:25:44 crc kubenswrapper[4778]: I0312 13:25:44.274432 4778 generic.go:334] "Generic (PLEG): container finished" podID="3418515c-3077-4237-9aa9-596ed9d3c137" containerID="33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a" exitCode=0 Mar 12 13:25:44 crc kubenswrapper[4778]: I0312 13:25:44.274480 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerDied","Data":"33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a"} Mar 12 13:25:45 crc kubenswrapper[4778]: I0312 13:25:45.279955 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerStarted","Data":"1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47"} Mar 12 13:25:45 crc kubenswrapper[4778]: I0312 13:25:45.295800 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pltxt" podStartSLOduration=1.707159239 podStartE2EDuration="4.295783543s" podCreationTimestamp="2026-03-12 13:25:41 +0000 UTC" firstStartedPulling="2026-03-12 13:25:42.2627421 +0000 UTC m=+960.711437496" lastFinishedPulling="2026-03-12 13:25:44.851366404 +0000 UTC m=+963.300061800" observedRunningTime="2026-03-12 13:25:45.294301731 +0000 UTC m=+963.742997137" watchObservedRunningTime="2026-03-12 13:25:45.295783543 +0000 UTC m=+963.744478939" Mar 12 13:25:47 crc kubenswrapper[4778]: I0312 13:25:47.363656 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68f5db54d6-zstmq" Mar 12 13:25:51 crc kubenswrapper[4778]: I0312 13:25:51.596064 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:51 crc kubenswrapper[4778]: I0312 13:25:51.596415 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:51 crc kubenswrapper[4778]: I0312 13:25:51.642558 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:52 crc kubenswrapper[4778]: I0312 13:25:52.657442 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:52 crc kubenswrapper[4778]: I0312 13:25:52.854373 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pltxt"] Mar 12 13:25:54 crc kubenswrapper[4778]: I0312 13:25:54.396794 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pltxt" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="registry-server" containerID="cri-o://1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47" gracePeriod=2 Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.282480 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.404791 4778 generic.go:334] "Generic (PLEG): container finished" podID="3418515c-3077-4237-9aa9-596ed9d3c137" containerID="1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47" exitCode=0 Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.404839 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerDied","Data":"1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47"} Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.404869 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pltxt" event={"ID":"3418515c-3077-4237-9aa9-596ed9d3c137","Type":"ContainerDied","Data":"b37770ad61325c6792ceeea5e8eb5ca55edd1557d07b2fc176c76ded639b7940"} Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.404889 4778 scope.go:117] "RemoveContainer" containerID="1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.405003 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pltxt" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.426583 4778 scope.go:117] "RemoveContainer" containerID="33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.472088 4778 scope.go:117] "RemoveContainer" containerID="6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.486625 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-utilities\") pod \"3418515c-3077-4237-9aa9-596ed9d3c137\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.486699 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b8rt\" (UniqueName: \"kubernetes.io/projected/3418515c-3077-4237-9aa9-596ed9d3c137-kube-api-access-2b8rt\") pod \"3418515c-3077-4237-9aa9-596ed9d3c137\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.486751 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-catalog-content\") pod \"3418515c-3077-4237-9aa9-596ed9d3c137\" (UID: \"3418515c-3077-4237-9aa9-596ed9d3c137\") " Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.495623 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3418515c-3077-4237-9aa9-596ed9d3c137-kube-api-access-2b8rt" (OuterVolumeSpecName: "kube-api-access-2b8rt") pod "3418515c-3077-4237-9aa9-596ed9d3c137" (UID: "3418515c-3077-4237-9aa9-596ed9d3c137"). InnerVolumeSpecName "kube-api-access-2b8rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.499026 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-utilities" (OuterVolumeSpecName: "utilities") pod "3418515c-3077-4237-9aa9-596ed9d3c137" (UID: "3418515c-3077-4237-9aa9-596ed9d3c137"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.508426 4778 scope.go:117] "RemoveContainer" containerID="1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47" Mar 12 13:25:55 crc kubenswrapper[4778]: E0312 13:25:55.509050 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47\": container with ID starting with 1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47 not found: ID does not exist" containerID="1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.509089 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47"} err="failed to get container status \"1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47\": rpc error: code = NotFound desc = could not find container \"1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47\": container with ID starting with 1d6d9105a852faf10ecfbf80e59f638b15c2f9a5383dbc2a9c8fb463b364dc47 not found: ID does not exist" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.509115 4778 scope.go:117] "RemoveContainer" containerID="33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a" Mar 12 13:25:55 crc kubenswrapper[4778]: E0312 13:25:55.509575 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a\": container with ID starting with 33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a not found: ID does not exist" containerID="33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.509613 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a"} err="failed to get container status \"33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a\": rpc error: code = NotFound desc = could not find container \"33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a\": container with ID starting with 33ac90bd1039472d943260eb6731eac265a3d0a77ac9f0a701175b768502761a not found: ID does not exist" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.509639 4778 scope.go:117] "RemoveContainer" containerID="6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864" Mar 12 13:25:55 crc kubenswrapper[4778]: E0312 13:25:55.509974 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864\": container with ID starting with 6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864 not found: ID does not exist" containerID="6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.510006 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864"} err="failed to get container status \"6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864\": rpc error: code = NotFound desc = could not find container \"6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864\": container with ID starting with 6b4907767b6af049385fd7daea9a886797c4e1193a43d026677035133dec2864 not found: ID does not exist" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.527253 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3418515c-3077-4237-9aa9-596ed9d3c137" (UID: "3418515c-3077-4237-9aa9-596ed9d3c137"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.587733 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.587764 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b8rt\" (UniqueName: \"kubernetes.io/projected/3418515c-3077-4237-9aa9-596ed9d3c137-kube-api-access-2b8rt\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.587774 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418515c-3077-4237-9aa9-596ed9d3c137-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.731008 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pltxt"] Mar 12 13:25:55 crc kubenswrapper[4778]: I0312 13:25:55.736980 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pltxt"] Mar 12 13:25:56 crc kubenswrapper[4778]: I0312 13:25:56.260957 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" path="/var/lib/kubelet/pods/3418515c-3077-4237-9aa9-596ed9d3c137/volumes" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.264873 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swx4c"] Mar 12 13:25:58 crc kubenswrapper[4778]: E0312 13:25:58.269271 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="registry-server" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.269299 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="registry-server" Mar 12 13:25:58 crc kubenswrapper[4778]: E0312 13:25:58.269318 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="extract-content" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.269330 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="extract-content" Mar 12 13:25:58 crc kubenswrapper[4778]: E0312 13:25:58.269347 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="extract-utilities" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.269355 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="extract-utilities" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.269475 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3418515c-3077-4237-9aa9-596ed9d3c137" containerName="registry-server" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.270487 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.290446 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swx4c"] Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.319061 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-utilities\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.319388 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n57qk\" (UniqueName: \"kubernetes.io/projected/19812411-eae6-4792-9f00-64a6604924fb-kube-api-access-n57qk\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.319523 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-catalog-content\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.420452 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-utilities\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.420515 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n57qk\" (UniqueName: \"kubernetes.io/projected/19812411-eae6-4792-9f00-64a6604924fb-kube-api-access-n57qk\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.420554 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-catalog-content\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.421054 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-catalog-content\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.421243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-utilities\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.449583 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n57qk\" (UniqueName: \"kubernetes.io/projected/19812411-eae6-4792-9f00-64a6604924fb-kube-api-access-n57qk\") pod \"community-operators-swx4c\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.558233 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.558289 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:25:58 crc kubenswrapper[4778]: I0312 13:25:58.593041 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:25:59 crc kubenswrapper[4778]: I0312 13:25:59.095757 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swx4c"] Mar 12 13:25:59 crc kubenswrapper[4778]: I0312 13:25:59.432772 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerStarted","Data":"4b8134073d8cbea56729606da6c6cdc39daa6d2a79458fae1697529ee25bbeab"} Mar 12 13:25:59 crc kubenswrapper[4778]: I0312 13:25:59.433106 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerStarted","Data":"044996b50b943738de55c5f921e61dd6fdf3f095e6a1a8170a40c8422279d3e8"} Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.130684 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555366-zt5bk"] Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.131401 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.133046 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.133307 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.134443 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.138987 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp4bm\" (UniqueName: \"kubernetes.io/projected/d48c598c-314b-4dc6-af90-7772a2ca7f2d-kube-api-access-jp4bm\") pod \"auto-csr-approver-29555366-zt5bk\" (UID: \"d48c598c-314b-4dc6-af90-7772a2ca7f2d\") " pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.140617 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555366-zt5bk"] Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.240112 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp4bm\" (UniqueName: \"kubernetes.io/projected/d48c598c-314b-4dc6-af90-7772a2ca7f2d-kube-api-access-jp4bm\") pod \"auto-csr-approver-29555366-zt5bk\" (UID: \"d48c598c-314b-4dc6-af90-7772a2ca7f2d\") " pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.258461 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp4bm\" (UniqueName: \"kubernetes.io/projected/d48c598c-314b-4dc6-af90-7772a2ca7f2d-kube-api-access-jp4bm\") pod \"auto-csr-approver-29555366-zt5bk\" (UID: \"d48c598c-314b-4dc6-af90-7772a2ca7f2d\") " pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.441040 4778 generic.go:334] "Generic (PLEG): container finished" podID="19812411-eae6-4792-9f00-64a6604924fb" containerID="4b8134073d8cbea56729606da6c6cdc39daa6d2a79458fae1697529ee25bbeab" exitCode=0 Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.441096 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerDied","Data":"4b8134073d8cbea56729606da6c6cdc39daa6d2a79458fae1697529ee25bbeab"} Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.448298 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:00 crc kubenswrapper[4778]: I0312 13:26:00.683958 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555366-zt5bk"] Mar 12 13:26:01 crc kubenswrapper[4778]: I0312 13:26:01.447996 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" event={"ID":"d48c598c-314b-4dc6-af90-7772a2ca7f2d","Type":"ContainerStarted","Data":"58b147806021f77dda3b41c28fbcd5bfb2f19e4aba276ad811e3cc8c42657ee6"} Mar 12 13:26:03 crc kubenswrapper[4778]: I0312 13:26:03.464287 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerStarted","Data":"c1bbf657f8b684a593ddef32aa60fde008a777e8a061554e91e9a2293d3a0832"} Mar 12 13:26:03 crc kubenswrapper[4778]: I0312 13:26:03.466816 4778 generic.go:334] "Generic (PLEG): container finished" podID="d48c598c-314b-4dc6-af90-7772a2ca7f2d" containerID="59816c72d24ee82ad1e212a580fdeb3c8cd671c1f79b421c31d995678ebec873" exitCode=0 Mar 12 13:26:03 crc kubenswrapper[4778]: I0312 13:26:03.466858 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" event={"ID":"d48c598c-314b-4dc6-af90-7772a2ca7f2d","Type":"ContainerDied","Data":"59816c72d24ee82ad1e212a580fdeb3c8cd671c1f79b421c31d995678ebec873"} Mar 12 13:26:04 crc kubenswrapper[4778]: I0312 13:26:04.474671 4778 generic.go:334] "Generic (PLEG): container finished" podID="19812411-eae6-4792-9f00-64a6604924fb" containerID="c1bbf657f8b684a593ddef32aa60fde008a777e8a061554e91e9a2293d3a0832" exitCode=0 Mar 12 13:26:04 crc kubenswrapper[4778]: I0312 13:26:04.474964 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerDied","Data":"c1bbf657f8b684a593ddef32aa60fde008a777e8a061554e91e9a2293d3a0832"} Mar 12 13:26:04 crc kubenswrapper[4778]: I0312 13:26:04.766034 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:04 crc kubenswrapper[4778]: I0312 13:26:04.814090 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp4bm\" (UniqueName: \"kubernetes.io/projected/d48c598c-314b-4dc6-af90-7772a2ca7f2d-kube-api-access-jp4bm\") pod \"d48c598c-314b-4dc6-af90-7772a2ca7f2d\" (UID: \"d48c598c-314b-4dc6-af90-7772a2ca7f2d\") " Mar 12 13:26:04 crc kubenswrapper[4778]: I0312 13:26:04.824740 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d48c598c-314b-4dc6-af90-7772a2ca7f2d-kube-api-access-jp4bm" (OuterVolumeSpecName: "kube-api-access-jp4bm") pod "d48c598c-314b-4dc6-af90-7772a2ca7f2d" (UID: "d48c598c-314b-4dc6-af90-7772a2ca7f2d"). InnerVolumeSpecName "kube-api-access-jp4bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:26:04 crc kubenswrapper[4778]: I0312 13:26:04.915645 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp4bm\" (UniqueName: \"kubernetes.io/projected/d48c598c-314b-4dc6-af90-7772a2ca7f2d-kube-api-access-jp4bm\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:05 crc kubenswrapper[4778]: I0312 13:26:05.483717 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" event={"ID":"d48c598c-314b-4dc6-af90-7772a2ca7f2d","Type":"ContainerDied","Data":"58b147806021f77dda3b41c28fbcd5bfb2f19e4aba276ad811e3cc8c42657ee6"} Mar 12 13:26:05 crc kubenswrapper[4778]: I0312 13:26:05.483761 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b147806021f77dda3b41c28fbcd5bfb2f19e4aba276ad811e3cc8c42657ee6" Mar 12 13:26:05 crc kubenswrapper[4778]: I0312 13:26:05.483836 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555366-zt5bk" Mar 12 13:26:05 crc kubenswrapper[4778]: I0312 13:26:05.834204 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555360-vwflx"] Mar 12 13:26:05 crc kubenswrapper[4778]: I0312 13:26:05.838848 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555360-vwflx"] Mar 12 13:26:06 crc kubenswrapper[4778]: I0312 13:26:06.261448 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c617404-7840-495c-80da-593af33f77d6" path="/var/lib/kubelet/pods/4c617404-7840-495c-80da-593af33f77d6/volumes" Mar 12 13:26:06 crc kubenswrapper[4778]: I0312 13:26:06.492499 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerStarted","Data":"6bbf817c355785c136024fac46b4f2a46cd308ddb02301de6e0f2fb81b7ff9b1"} Mar 12 13:26:06 crc kubenswrapper[4778]: I0312 13:26:06.517160 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swx4c" podStartSLOduration=2.65187224 podStartE2EDuration="8.517137619s" podCreationTimestamp="2026-03-12 13:25:58 +0000 UTC" firstStartedPulling="2026-03-12 13:26:00.443721841 +0000 UTC m=+978.892417237" lastFinishedPulling="2026-03-12 13:26:06.3089871 +0000 UTC m=+984.757682616" observedRunningTime="2026-03-12 13:26:06.513611578 +0000 UTC m=+984.962306984" watchObservedRunningTime="2026-03-12 13:26:06.517137619 +0000 UTC m=+984.965833015" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.052905 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-54d5c4b6c7-gh4lx" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.495942 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5jsf9"] Mar 12 13:26:07 crc kubenswrapper[4778]: E0312 13:26:07.496703 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48c598c-314b-4dc6-af90-7772a2ca7f2d" containerName="oc" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.496726 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48c598c-314b-4dc6-af90-7772a2ca7f2d" containerName="oc" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.496899 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48c598c-314b-4dc6-af90-7772a2ca7f2d" containerName="oc" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.497905 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.506241 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5jsf9"] Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.574371 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-utilities\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.574423 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-catalog-content\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.574459 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfw95\" (UniqueName: \"kubernetes.io/projected/631910e5-eefd-4ccf-adde-4609f7825e27-kube-api-access-qfw95\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.675779 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfw95\" (UniqueName: \"kubernetes.io/projected/631910e5-eefd-4ccf-adde-4609f7825e27-kube-api-access-qfw95\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.675884 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-utilities\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.675904 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-catalog-content\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.676375 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-catalog-content\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.676797 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-utilities\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.699786 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfw95\" (UniqueName: \"kubernetes.io/projected/631910e5-eefd-4ccf-adde-4609f7825e27-kube-api-access-qfw95\") pod \"certified-operators-5jsf9\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.854654 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.918745 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f"] Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.919513 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.923981 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-r8pt2" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.924162 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.953957 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zxv5p"] Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.956788 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.959425 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.964008 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.967624 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f"] Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.997740 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-k7nvk"] Mar 12 13:26:07 crc kubenswrapper[4778]: I0312 13:26:07.998810 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.007099 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.007284 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.007462 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.007538 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-d5gxv" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.009394 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-mnjql"] Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.010730 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.012128 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.055834 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-mnjql"] Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056399 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-conf\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056426 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-sockets\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056455 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-startup\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056500 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056520 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-reloader\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056544 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9ns2\" (UniqueName: \"kubernetes.io/projected/2f214887-d638-42fa-aa86-1518cfae600d-kube-api-access-q9ns2\") pod \"frr-k8s-webhook-server-bcc4b6f68-x2n7f\" (UID: \"2f214887-d638-42fa-aa86-1518cfae600d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056565 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w8tl\" (UniqueName: \"kubernetes.io/projected/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-kube-api-access-2w8tl\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056608 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics-certs\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.056634 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2f214887-d638-42fa-aa86-1518cfae600d-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-x2n7f\" (UID: \"2f214887-d638-42fa-aa86-1518cfae600d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157425 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9ns2\" (UniqueName: \"kubernetes.io/projected/2f214887-d638-42fa-aa86-1518cfae600d-kube-api-access-q9ns2\") pod \"frr-k8s-webhook-server-bcc4b6f68-x2n7f\" (UID: \"2f214887-d638-42fa-aa86-1518cfae600d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157488 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w8tl\" (UniqueName: \"kubernetes.io/projected/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-kube-api-access-2w8tl\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157521 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics-certs\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157547 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metrics-certs\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157582 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2f214887-d638-42fa-aa86-1518cfae600d-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-x2n7f\" (UID: \"2f214887-d638-42fa-aa86-1518cfae600d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157620 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157647 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-conf\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157669 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/14351deb-3286-4464-8eac-6bb116a9ebce-metrics-certs\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157690 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-sockets\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157726 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-startup\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157760 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9tqj\" (UniqueName: \"kubernetes.io/projected/14351deb-3286-4464-8eac-6bb116a9ebce-kube-api-access-c9tqj\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157809 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzjv2\" (UniqueName: \"kubernetes.io/projected/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-kube-api-access-kzjv2\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157837 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157858 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-reloader\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157882 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metallb-excludel2\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.157903 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14351deb-3286-4464-8eac-6bb116a9ebce-cert\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.158399 4778 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.158445 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics-certs podName:b5f035ed-2e64-4000-908f-6d0ecab1fe8d nodeName:}" failed. No retries permitted until 2026-03-12 13:26:08.658427656 +0000 UTC m=+987.107123052 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics-certs") pod "frr-k8s-zxv5p" (UID: "b5f035ed-2e64-4000-908f-6d0ecab1fe8d") : secret "frr-k8s-certs-secret" not found Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.158942 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-sockets\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.159735 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-startup\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.160069 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.160345 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-reloader\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.160530 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-frr-conf\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.166884 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2f214887-d638-42fa-aa86-1518cfae600d-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-x2n7f\" (UID: \"2f214887-d638-42fa-aa86-1518cfae600d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.177773 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9ns2\" (UniqueName: \"kubernetes.io/projected/2f214887-d638-42fa-aa86-1518cfae600d-kube-api-access-q9ns2\") pod \"frr-k8s-webhook-server-bcc4b6f68-x2n7f\" (UID: \"2f214887-d638-42fa-aa86-1518cfae600d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.189264 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w8tl\" (UniqueName: \"kubernetes.io/projected/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-kube-api-access-2w8tl\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.246633 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.260984 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzjv2\" (UniqueName: \"kubernetes.io/projected/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-kube-api-access-kzjv2\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.261040 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metallb-excludel2\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.261063 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14351deb-3286-4464-8eac-6bb116a9ebce-cert\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.261121 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metrics-certs\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.261150 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.261192 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/14351deb-3286-4464-8eac-6bb116a9ebce-metrics-certs\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.261248 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9tqj\" (UniqueName: \"kubernetes.io/projected/14351deb-3286-4464-8eac-6bb116a9ebce-kube-api-access-c9tqj\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.262824 4778 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.262834 4778 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.262867 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist podName:f2e1d11e-8f27-498d-8d45-ac0e14a796fe nodeName:}" failed. No retries permitted until 2026-03-12 13:26:08.762853785 +0000 UTC m=+987.211549181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist") pod "speaker-k7nvk" (UID: "f2e1d11e-8f27-498d-8d45-ac0e14a796fe") : secret "metallb-memberlist" not found Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.262928 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metrics-certs podName:f2e1d11e-8f27-498d-8d45-ac0e14a796fe nodeName:}" failed. No retries permitted until 2026-03-12 13:26:08.762904937 +0000 UTC m=+987.211600403 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metrics-certs") pod "speaker-k7nvk" (UID: "f2e1d11e-8f27-498d-8d45-ac0e14a796fe") : secret "speaker-certs-secret" not found Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.263995 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metallb-excludel2\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.276711 4778 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.277649 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/14351deb-3286-4464-8eac-6bb116a9ebce-metrics-certs\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.304708 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzjv2\" (UniqueName: \"kubernetes.io/projected/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-kube-api-access-kzjv2\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.306555 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14351deb-3286-4464-8eac-6bb116a9ebce-cert\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.311916 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9tqj\" (UniqueName: \"kubernetes.io/projected/14351deb-3286-4464-8eac-6bb116a9ebce-kube-api-access-c9tqj\") pod \"controller-7bb4cc7c98-mnjql\" (UID: \"14351deb-3286-4464-8eac-6bb116a9ebce\") " pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.337512 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.380420 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5jsf9"] Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.594017 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.594064 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.668486 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics-certs\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.673423 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f035ed-2e64-4000-908f-6d0ecab1fe8d-metrics-certs\") pod \"frr-k8s-zxv5p\" (UID: \"b5f035ed-2e64-4000-908f-6d0ecab1fe8d\") " pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.679607 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerStarted","Data":"62dfee05353fcfd0dfc997669714bab327d0be23b5ebdac66ccce3caa2514ce0"} Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.783732 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metrics-certs\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.783807 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.783948 4778 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 12 13:26:08 crc kubenswrapper[4778]: E0312 13:26:08.784001 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist podName:f2e1d11e-8f27-498d-8d45-ac0e14a796fe nodeName:}" failed. No retries permitted until 2026-03-12 13:26:09.783984914 +0000 UTC m=+988.232680310 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist") pod "speaker-k7nvk" (UID: "f2e1d11e-8f27-498d-8d45-ac0e14a796fe") : secret "metallb-memberlist" not found Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.805915 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-metrics-certs\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.835500 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f"] Mar 12 13:26:08 crc kubenswrapper[4778]: I0312 13:26:08.883313 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.085658 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-mnjql"] Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.690266 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-mnjql" event={"ID":"14351deb-3286-4464-8eac-6bb116a9ebce","Type":"ContainerStarted","Data":"208f514cd919cc1b600edb4d8ee8454e19627f040e41b48e238f7a46fdcddc04"} Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.690331 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-mnjql" event={"ID":"14351deb-3286-4464-8eac-6bb116a9ebce","Type":"ContainerStarted","Data":"65381005f96f5204bdfdae7f78e7a3fa233be0b67d02a80bc373154ecb874d3d"} Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.690344 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-mnjql" event={"ID":"14351deb-3286-4464-8eac-6bb116a9ebce","Type":"ContainerStarted","Data":"5032f44852cb1aa63127bd5fc1273012f6f299fadd0f65320f81ac5a2288de17"} Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.690637 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.694239 4778 generic.go:334] "Generic (PLEG): container finished" podID="631910e5-eefd-4ccf-adde-4609f7825e27" containerID="2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668" exitCode=0 Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.694326 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerDied","Data":"2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668"} Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.696889 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" event={"ID":"2f214887-d638-42fa-aa86-1518cfae600d","Type":"ContainerStarted","Data":"fa24e6fcabd1931dbd69e0147a2ed76a5df3ab1f219dd4e2b4b1a4991f832bdc"} Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.711757 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"0e9d2b811f145ddb3e597854983500d583dee7155a4d2d7675423529527972d6"} Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.719514 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-mnjql" podStartSLOduration=2.719491313 podStartE2EDuration="2.719491313s" podCreationTimestamp="2026-03-12 13:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:26:09.717697192 +0000 UTC m=+988.166392598" watchObservedRunningTime="2026-03-12 13:26:09.719491313 +0000 UTC m=+988.168186709" Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.748013 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-swx4c" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="registry-server" probeResult="failure" output=< Mar 12 13:26:09 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:26:09 crc kubenswrapper[4778]: > Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.799774 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.813802 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f2e1d11e-8f27-498d-8d45-ac0e14a796fe-memberlist\") pod \"speaker-k7nvk\" (UID: \"f2e1d11e-8f27-498d-8d45-ac0e14a796fe\") " pod="metallb-system/speaker-k7nvk" Mar 12 13:26:09 crc kubenswrapper[4778]: I0312 13:26:09.824751 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-k7nvk" Mar 12 13:26:09 crc kubenswrapper[4778]: W0312 13:26:09.848306 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2e1d11e_8f27_498d_8d45_ac0e14a796fe.slice/crio-12888157619e57ff85ec07c59ba39825a38220466903e20336f50d93c86e2f8c WatchSource:0}: Error finding container 12888157619e57ff85ec07c59ba39825a38220466903e20336f50d93c86e2f8c: Status 404 returned error can't find the container with id 12888157619e57ff85ec07c59ba39825a38220466903e20336f50d93c86e2f8c Mar 12 13:26:10 crc kubenswrapper[4778]: I0312 13:26:10.724795 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-k7nvk" event={"ID":"f2e1d11e-8f27-498d-8d45-ac0e14a796fe","Type":"ContainerStarted","Data":"7ca9ae77fb3b5fe15347f655f81545506968ffe60f16f3351a6eae47c6693bd7"} Mar 12 13:26:10 crc kubenswrapper[4778]: I0312 13:26:10.725260 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-k7nvk" event={"ID":"f2e1d11e-8f27-498d-8d45-ac0e14a796fe","Type":"ContainerStarted","Data":"3826db80e29ade0a6f881d928d89599e164f8e6265d83b75f8c9c37cdf911faa"} Mar 12 13:26:10 crc kubenswrapper[4778]: I0312 13:26:10.725274 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-k7nvk" event={"ID":"f2e1d11e-8f27-498d-8d45-ac0e14a796fe","Type":"ContainerStarted","Data":"12888157619e57ff85ec07c59ba39825a38220466903e20336f50d93c86e2f8c"} Mar 12 13:26:10 crc kubenswrapper[4778]: I0312 13:26:10.725918 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-k7nvk" Mar 12 13:26:10 crc kubenswrapper[4778]: I0312 13:26:10.754911 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-k7nvk" podStartSLOduration=3.754896134 podStartE2EDuration="3.754896134s" podCreationTimestamp="2026-03-12 13:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:26:10.753993699 +0000 UTC m=+989.202689095" watchObservedRunningTime="2026-03-12 13:26:10.754896134 +0000 UTC m=+989.203591530" Mar 12 13:26:11 crc kubenswrapper[4778]: I0312 13:26:11.905824 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerStarted","Data":"21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96"} Mar 12 13:26:13 crc kubenswrapper[4778]: E0312 13:26:13.478533 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631910e5_eefd_4ccf_adde_4609f7825e27.slice/crio-conmon-21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96.scope\": RecentStats: unable to find data in memory cache]" Mar 12 13:26:13 crc kubenswrapper[4778]: I0312 13:26:13.925740 4778 generic.go:334] "Generic (PLEG): container finished" podID="631910e5-eefd-4ccf-adde-4609f7825e27" containerID="21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96" exitCode=0 Mar 12 13:26:13 crc kubenswrapper[4778]: I0312 13:26:13.925787 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerDied","Data":"21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96"} Mar 12 13:26:15 crc kubenswrapper[4778]: I0312 13:26:15.225796 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerStarted","Data":"cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135"} Mar 12 13:26:15 crc kubenswrapper[4778]: I0312 13:26:15.249440 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5jsf9" podStartSLOduration=3.213368395 podStartE2EDuration="8.249406626s" podCreationTimestamp="2026-03-12 13:26:07 +0000 UTC" firstStartedPulling="2026-03-12 13:26:09.69588754 +0000 UTC m=+988.144582956" lastFinishedPulling="2026-03-12 13:26:14.731925801 +0000 UTC m=+993.180621187" observedRunningTime="2026-03-12 13:26:15.248567232 +0000 UTC m=+993.697262658" watchObservedRunningTime="2026-03-12 13:26:15.249406626 +0000 UTC m=+993.698102022" Mar 12 13:26:17 crc kubenswrapper[4778]: I0312 13:26:17.855781 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:17 crc kubenswrapper[4778]: I0312 13:26:17.856691 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:18 crc kubenswrapper[4778]: I0312 13:26:18.683719 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:26:18 crc kubenswrapper[4778]: I0312 13:26:18.742038 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:26:18 crc kubenswrapper[4778]: I0312 13:26:18.930865 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swx4c"] Mar 12 13:26:19 crc kubenswrapper[4778]: I0312 13:26:19.012604 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5jsf9" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="registry-server" probeResult="failure" output=< Mar 12 13:26:19 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:26:19 crc kubenswrapper[4778]: > Mar 12 13:26:20 crc kubenswrapper[4778]: I0312 13:26:20.384237 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swx4c" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="registry-server" containerID="cri-o://6bbf817c355785c136024fac46b4f2a46cd308ddb02301de6e0f2fb81b7ff9b1" gracePeriod=2 Mar 12 13:26:21 crc kubenswrapper[4778]: I0312 13:26:21.393094 4778 generic.go:334] "Generic (PLEG): container finished" podID="19812411-eae6-4792-9f00-64a6604924fb" containerID="6bbf817c355785c136024fac46b4f2a46cd308ddb02301de6e0f2fb81b7ff9b1" exitCode=0 Mar 12 13:26:21 crc kubenswrapper[4778]: I0312 13:26:21.393131 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerDied","Data":"6bbf817c355785c136024fac46b4f2a46cd308ddb02301de6e0f2fb81b7ff9b1"} Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.799296 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.816866 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-catalog-content\") pod \"19812411-eae6-4792-9f00-64a6604924fb\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.816973 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-utilities\") pod \"19812411-eae6-4792-9f00-64a6604924fb\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.817007 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n57qk\" (UniqueName: \"kubernetes.io/projected/19812411-eae6-4792-9f00-64a6604924fb-kube-api-access-n57qk\") pod \"19812411-eae6-4792-9f00-64a6604924fb\" (UID: \"19812411-eae6-4792-9f00-64a6604924fb\") " Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.821334 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-utilities" (OuterVolumeSpecName: "utilities") pod "19812411-eae6-4792-9f00-64a6604924fb" (UID: "19812411-eae6-4792-9f00-64a6604924fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.831296 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19812411-eae6-4792-9f00-64a6604924fb-kube-api-access-n57qk" (OuterVolumeSpecName: "kube-api-access-n57qk") pod "19812411-eae6-4792-9f00-64a6604924fb" (UID: "19812411-eae6-4792-9f00-64a6604924fb"). InnerVolumeSpecName "kube-api-access-n57qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.890744 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19812411-eae6-4792-9f00-64a6604924fb" (UID: "19812411-eae6-4792-9f00-64a6604924fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.920159 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.920822 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19812411-eae6-4792-9f00-64a6604924fb-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:22 crc kubenswrapper[4778]: I0312 13:26:22.920834 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n57qk\" (UniqueName: \"kubernetes.io/projected/19812411-eae6-4792-9f00-64a6604924fb-kube-api-access-n57qk\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.407762 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" event={"ID":"2f214887-d638-42fa-aa86-1518cfae600d","Type":"ContainerStarted","Data":"2c4c6771d80567c3899e80a38ef6f909606c5c92cc88ac51b59fc3e170d8c825"} Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.408172 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.409242 4778 generic.go:334] "Generic (PLEG): container finished" podID="b5f035ed-2e64-4000-908f-6d0ecab1fe8d" containerID="4c6923df2cdecc82edcf2736d86f11607a489acc252c64dfbe59ae1943318eb0" exitCode=0 Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.409317 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerDied","Data":"4c6923df2cdecc82edcf2736d86f11607a489acc252c64dfbe59ae1943318eb0"} Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.411592 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swx4c" event={"ID":"19812411-eae6-4792-9f00-64a6604924fb","Type":"ContainerDied","Data":"044996b50b943738de55c5f921e61dd6fdf3f095e6a1a8170a40c8422279d3e8"} Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.411649 4778 scope.go:117] "RemoveContainer" containerID="6bbf817c355785c136024fac46b4f2a46cd308ddb02301de6e0f2fb81b7ff9b1" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.411661 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swx4c" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.431919 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" podStartSLOduration=2.4764062989999998 podStartE2EDuration="16.431898598s" podCreationTimestamp="2026-03-12 13:26:07 +0000 UTC" firstStartedPulling="2026-03-12 13:26:08.889473283 +0000 UTC m=+987.338168679" lastFinishedPulling="2026-03-12 13:26:22.844965582 +0000 UTC m=+1001.293660978" observedRunningTime="2026-03-12 13:26:23.430307742 +0000 UTC m=+1001.879003158" watchObservedRunningTime="2026-03-12 13:26:23.431898598 +0000 UTC m=+1001.880594004" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.488301 4778 scope.go:117] "RemoveContainer" containerID="c1bbf657f8b684a593ddef32aa60fde008a777e8a061554e91e9a2293d3a0832" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.527338 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swx4c"] Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.528067 4778 scope.go:117] "RemoveContainer" containerID="4b8134073d8cbea56729606da6c6cdc39daa6d2a79458fae1697529ee25bbeab" Mar 12 13:26:23 crc kubenswrapper[4778]: I0312 13:26:23.533277 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swx4c"] Mar 12 13:26:23 crc kubenswrapper[4778]: E0312 13:26:23.650651 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19812411_eae6_4792_9f00_64a6604924fb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19812411_eae6_4792_9f00_64a6604924fb.slice/crio-044996b50b943738de55c5f921e61dd6fdf3f095e6a1a8170a40c8422279d3e8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5f035ed_2e64_4000_908f_6d0ecab1fe8d.slice/crio-conmon-5f944b02e5c4fd2938ea1f2eed17a7cf148a1ac57c7cf02afb4110bb86aa032a.scope\": RecentStats: unable to find data in memory cache]" Mar 12 13:26:24 crc kubenswrapper[4778]: I0312 13:26:24.268102 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19812411-eae6-4792-9f00-64a6604924fb" path="/var/lib/kubelet/pods/19812411-eae6-4792-9f00-64a6604924fb/volumes" Mar 12 13:26:24 crc kubenswrapper[4778]: I0312 13:26:24.423508 4778 generic.go:334] "Generic (PLEG): container finished" podID="b5f035ed-2e64-4000-908f-6d0ecab1fe8d" containerID="5f944b02e5c4fd2938ea1f2eed17a7cf148a1ac57c7cf02afb4110bb86aa032a" exitCode=0 Mar 12 13:26:24 crc kubenswrapper[4778]: I0312 13:26:24.423560 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerDied","Data":"5f944b02e5c4fd2938ea1f2eed17a7cf148a1ac57c7cf02afb4110bb86aa032a"} Mar 12 13:26:25 crc kubenswrapper[4778]: I0312 13:26:25.433769 4778 generic.go:334] "Generic (PLEG): container finished" podID="b5f035ed-2e64-4000-908f-6d0ecab1fe8d" containerID="d4b23df880c826685e70ef1ea6f346e4940a8db232ba57d0a989bdee1e47c504" exitCode=0 Mar 12 13:26:25 crc kubenswrapper[4778]: I0312 13:26:25.433871 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerDied","Data":"d4b23df880c826685e70ef1ea6f346e4940a8db232ba57d0a989bdee1e47c504"} Mar 12 13:26:26 crc kubenswrapper[4778]: I0312 13:26:26.443903 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"69d841216da693033b93967bd9f96f5622fe3a6eaaca22f4423b12fc72860396"} Mar 12 13:26:26 crc kubenswrapper[4778]: I0312 13:26:26.443944 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"b8c78d004bac3227097fce8d27ac831d2de1d86df94ffb03d839bc1cf5623403"} Mar 12 13:26:26 crc kubenswrapper[4778]: I0312 13:26:26.443954 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"7cb1446a7230327ccf23a0d317d5a79825f8d6e0075315bff24964579f60572e"} Mar 12 13:26:26 crc kubenswrapper[4778]: I0312 13:26:26.443964 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"36f2dfd7ce0a51b4cd99ee0a1580c67be8228e9de0375a5e2096190edb364201"} Mar 12 13:26:26 crc kubenswrapper[4778]: I0312 13:26:26.443972 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"61a03e1c5b6356ee74a455f3f4a569a67017490c8228f33d9eecbb641c5896fe"} Mar 12 13:26:27 crc kubenswrapper[4778]: I0312 13:26:27.454482 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zxv5p" event={"ID":"b5f035ed-2e64-4000-908f-6d0ecab1fe8d","Type":"ContainerStarted","Data":"8625916bc5f9b7f34622f5f38a1d5e954161f0390eea62087f4ed0fea0b61d44"} Mar 12 13:26:27 crc kubenswrapper[4778]: I0312 13:26:27.455116 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:27 crc kubenswrapper[4778]: I0312 13:26:27.480247 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zxv5p" podStartSLOduration=6.738395297 podStartE2EDuration="20.480228699s" podCreationTimestamp="2026-03-12 13:26:07 +0000 UTC" firstStartedPulling="2026-03-12 13:26:09.069637534 +0000 UTC m=+987.518332930" lastFinishedPulling="2026-03-12 13:26:22.811470936 +0000 UTC m=+1001.260166332" observedRunningTime="2026-03-12 13:26:27.476559345 +0000 UTC m=+1005.925254751" watchObservedRunningTime="2026-03-12 13:26:27.480228699 +0000 UTC m=+1005.928924105" Mar 12 13:26:27 crc kubenswrapper[4778]: I0312 13:26:27.919098 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:27 crc kubenswrapper[4778]: I0312 13:26:27.968166 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.154205 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5jsf9"] Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.342123 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-mnjql" Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.558313 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.558406 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.558467 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.559363 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b65e287d42eea6146877a35b0789c26ac0ef9f5d251a760b59f08b3fef055d65"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.559455 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://b65e287d42eea6146877a35b0789c26ac0ef9f5d251a760b59f08b3fef055d65" gracePeriod=600 Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.884223 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:28 crc kubenswrapper[4778]: I0312 13:26:28.928404 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:29 crc kubenswrapper[4778]: I0312 13:26:29.467679 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="b65e287d42eea6146877a35b0789c26ac0ef9f5d251a760b59f08b3fef055d65" exitCode=0 Mar 12 13:26:29 crc kubenswrapper[4778]: I0312 13:26:29.467765 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"b65e287d42eea6146877a35b0789c26ac0ef9f5d251a760b59f08b3fef055d65"} Mar 12 13:26:29 crc kubenswrapper[4778]: I0312 13:26:29.467810 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"3b4b372cac8f288fc2585670d5ab7c00c41331f173130d39b164aa74e4e3e398"} Mar 12 13:26:29 crc kubenswrapper[4778]: I0312 13:26:29.467826 4778 scope.go:117] "RemoveContainer" containerID="dfcc37339849724c4aacca3262255dd43897a2284c2172380a90cc97f52e3a46" Mar 12 13:26:29 crc kubenswrapper[4778]: I0312 13:26:29.468059 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5jsf9" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="registry-server" containerID="cri-o://cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135" gracePeriod=2 Mar 12 13:26:29 crc kubenswrapper[4778]: I0312 13:26:29.828626 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-k7nvk" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.356378 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.436620 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-utilities\") pod \"631910e5-eefd-4ccf-adde-4609f7825e27\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.437142 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfw95\" (UniqueName: \"kubernetes.io/projected/631910e5-eefd-4ccf-adde-4609f7825e27-kube-api-access-qfw95\") pod \"631910e5-eefd-4ccf-adde-4609f7825e27\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.437191 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-catalog-content\") pod \"631910e5-eefd-4ccf-adde-4609f7825e27\" (UID: \"631910e5-eefd-4ccf-adde-4609f7825e27\") " Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.437953 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-utilities" (OuterVolumeSpecName: "utilities") pod "631910e5-eefd-4ccf-adde-4609f7825e27" (UID: "631910e5-eefd-4ccf-adde-4609f7825e27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.444480 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631910e5-eefd-4ccf-adde-4609f7825e27-kube-api-access-qfw95" (OuterVolumeSpecName: "kube-api-access-qfw95") pod "631910e5-eefd-4ccf-adde-4609f7825e27" (UID: "631910e5-eefd-4ccf-adde-4609f7825e27"). InnerVolumeSpecName "kube-api-access-qfw95". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.476260 4778 generic.go:334] "Generic (PLEG): container finished" podID="631910e5-eefd-4ccf-adde-4609f7825e27" containerID="cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135" exitCode=0 Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.476325 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jsf9" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.476342 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerDied","Data":"cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135"} Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.476380 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jsf9" event={"ID":"631910e5-eefd-4ccf-adde-4609f7825e27","Type":"ContainerDied","Data":"62dfee05353fcfd0dfc997669714bab327d0be23b5ebdac66ccce3caa2514ce0"} Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.476402 4778 scope.go:117] "RemoveContainer" containerID="cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.497952 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "631910e5-eefd-4ccf-adde-4609f7825e27" (UID: "631910e5-eefd-4ccf-adde-4609f7825e27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.502181 4778 scope.go:117] "RemoveContainer" containerID="21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.517033 4778 scope.go:117] "RemoveContainer" containerID="2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.538261 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.538287 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfw95\" (UniqueName: \"kubernetes.io/projected/631910e5-eefd-4ccf-adde-4609f7825e27-kube-api-access-qfw95\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.538298 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631910e5-eefd-4ccf-adde-4609f7825e27-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.538694 4778 scope.go:117] "RemoveContainer" containerID="cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135" Mar 12 13:26:30 crc kubenswrapper[4778]: E0312 13:26:30.539723 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135\": container with ID starting with cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135 not found: ID does not exist" containerID="cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.539755 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135"} err="failed to get container status \"cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135\": rpc error: code = NotFound desc = could not find container \"cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135\": container with ID starting with cbb6037b8cc080048e6578c32ace5294078d9b4d705e389bb202aad65ec4b135 not found: ID does not exist" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.539778 4778 scope.go:117] "RemoveContainer" containerID="21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96" Mar 12 13:26:30 crc kubenswrapper[4778]: E0312 13:26:30.540360 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96\": container with ID starting with 21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96 not found: ID does not exist" containerID="21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.540433 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96"} err="failed to get container status \"21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96\": rpc error: code = NotFound desc = could not find container \"21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96\": container with ID starting with 21b20ff1eb5f0382d62c5bf4557691c60f975b3c3036f939f288f19834d29b96 not found: ID does not exist" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.540536 4778 scope.go:117] "RemoveContainer" containerID="2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668" Mar 12 13:26:30 crc kubenswrapper[4778]: E0312 13:26:30.542713 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668\": container with ID starting with 2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668 not found: ID does not exist" containerID="2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.542747 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668"} err="failed to get container status \"2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668\": rpc error: code = NotFound desc = could not find container \"2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668\": container with ID starting with 2379ded140d6beb0318ea43c0d651b20477999f6fc28a1039221cf9aa630c668 not found: ID does not exist" Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.809363 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5jsf9"] Mar 12 13:26:30 crc kubenswrapper[4778]: I0312 13:26:30.816229 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5jsf9"] Mar 12 13:26:32 crc kubenswrapper[4778]: I0312 13:26:32.260535 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" path="/var/lib/kubelet/pods/631910e5-eefd-4ccf-adde-4609f7825e27/volumes" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561088 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tbh2n"] Mar 12 13:26:34 crc kubenswrapper[4778]: E0312 13:26:34.561559 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="extract-utilities" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561570 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="extract-utilities" Mar 12 13:26:34 crc kubenswrapper[4778]: E0312 13:26:34.561579 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="registry-server" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561586 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="registry-server" Mar 12 13:26:34 crc kubenswrapper[4778]: E0312 13:26:34.561597 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="extract-content" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561603 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="extract-content" Mar 12 13:26:34 crc kubenswrapper[4778]: E0312 13:26:34.561614 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="extract-content" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561619 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="extract-content" Mar 12 13:26:34 crc kubenswrapper[4778]: E0312 13:26:34.561630 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="extract-utilities" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561635 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="extract-utilities" Mar 12 13:26:34 crc kubenswrapper[4778]: E0312 13:26:34.561642 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="registry-server" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561647 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="registry-server" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561756 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="19812411-eae6-4792-9f00-64a6604924fb" containerName="registry-server" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.561774 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="631910e5-eefd-4ccf-adde-4609f7825e27" containerName="registry-server" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.562118 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.563963 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-dl4ht" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.564618 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.565242 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.575963 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tbh2n"] Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.583164 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvdhc\" (UniqueName: \"kubernetes.io/projected/60754672-3d3c-4763-8078-356d0a0167ac-kube-api-access-gvdhc\") pod \"openstack-operator-index-tbh2n\" (UID: \"60754672-3d3c-4763-8078-356d0a0167ac\") " pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.684493 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvdhc\" (UniqueName: \"kubernetes.io/projected/60754672-3d3c-4763-8078-356d0a0167ac-kube-api-access-gvdhc\") pod \"openstack-operator-index-tbh2n\" (UID: \"60754672-3d3c-4763-8078-356d0a0167ac\") " pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.708739 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvdhc\" (UniqueName: \"kubernetes.io/projected/60754672-3d3c-4763-8078-356d0a0167ac-kube-api-access-gvdhc\") pod \"openstack-operator-index-tbh2n\" (UID: \"60754672-3d3c-4763-8078-356d0a0167ac\") " pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:34 crc kubenswrapper[4778]: I0312 13:26:34.880894 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:35 crc kubenswrapper[4778]: I0312 13:26:35.312636 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tbh2n"] Mar 12 13:26:35 crc kubenswrapper[4778]: I0312 13:26:35.513762 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tbh2n" event={"ID":"60754672-3d3c-4763-8078-356d0a0167ac","Type":"ContainerStarted","Data":"82b0ae343a0a10b9762a77419cf9048e88f45587940a8f410652a383e6167520"} Mar 12 13:26:38 crc kubenswrapper[4778]: I0312 13:26:38.251969 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-x2n7f" Mar 12 13:26:38 crc kubenswrapper[4778]: I0312 13:26:38.537431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tbh2n" event={"ID":"60754672-3d3c-4763-8078-356d0a0167ac","Type":"ContainerStarted","Data":"b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a"} Mar 12 13:26:38 crc kubenswrapper[4778]: I0312 13:26:38.889004 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zxv5p" Mar 12 13:26:38 crc kubenswrapper[4778]: I0312 13:26:38.925523 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tbh2n" podStartSLOduration=2.789810305 podStartE2EDuration="4.925496408s" podCreationTimestamp="2026-03-12 13:26:34 +0000 UTC" firstStartedPulling="2026-03-12 13:26:35.321596917 +0000 UTC m=+1013.770292313" lastFinishedPulling="2026-03-12 13:26:37.45728301 +0000 UTC m=+1015.905978416" observedRunningTime="2026-03-12 13:26:38.555472041 +0000 UTC m=+1017.004167437" watchObservedRunningTime="2026-03-12 13:26:38.925496408 +0000 UTC m=+1017.374191834" Mar 12 13:26:38 crc kubenswrapper[4778]: I0312 13:26:38.953546 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-tbh2n"] Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.566428 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-b2fsv"] Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.567561 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.576582 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-b2fsv"] Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.652566 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk5sj\" (UniqueName: \"kubernetes.io/projected/748546a6-1355-470f-b8d0-de395cf3f681-kube-api-access-tk5sj\") pod \"openstack-operator-index-b2fsv\" (UID: \"748546a6-1355-470f-b8d0-de395cf3f681\") " pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.753564 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk5sj\" (UniqueName: \"kubernetes.io/projected/748546a6-1355-470f-b8d0-de395cf3f681-kube-api-access-tk5sj\") pod \"openstack-operator-index-b2fsv\" (UID: \"748546a6-1355-470f-b8d0-de395cf3f681\") " pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.771412 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk5sj\" (UniqueName: \"kubernetes.io/projected/748546a6-1355-470f-b8d0-de395cf3f681-kube-api-access-tk5sj\") pod \"openstack-operator-index-b2fsv\" (UID: \"748546a6-1355-470f-b8d0-de395cf3f681\") " pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:39 crc kubenswrapper[4778]: I0312 13:26:39.886149 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:40 crc kubenswrapper[4778]: I0312 13:26:40.323525 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-b2fsv"] Mar 12 13:26:40 crc kubenswrapper[4778]: W0312 13:26:40.334422 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod748546a6_1355_470f_b8d0_de395cf3f681.slice/crio-eecc88b48278b2739e6f24ef7a7358771a2419e4c79e671aa9590dbed5f25c75 WatchSource:0}: Error finding container eecc88b48278b2739e6f24ef7a7358771a2419e4c79e671aa9590dbed5f25c75: Status 404 returned error can't find the container with id eecc88b48278b2739e6f24ef7a7358771a2419e4c79e671aa9590dbed5f25c75 Mar 12 13:26:40 crc kubenswrapper[4778]: I0312 13:26:40.551011 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-b2fsv" event={"ID":"748546a6-1355-470f-b8d0-de395cf3f681","Type":"ContainerStarted","Data":"eecc88b48278b2739e6f24ef7a7358771a2419e4c79e671aa9590dbed5f25c75"} Mar 12 13:26:40 crc kubenswrapper[4778]: I0312 13:26:40.551149 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-tbh2n" podUID="60754672-3d3c-4763-8078-356d0a0167ac" containerName="registry-server" containerID="cri-o://b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a" gracePeriod=2 Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.003544 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.171943 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvdhc\" (UniqueName: \"kubernetes.io/projected/60754672-3d3c-4763-8078-356d0a0167ac-kube-api-access-gvdhc\") pod \"60754672-3d3c-4763-8078-356d0a0167ac\" (UID: \"60754672-3d3c-4763-8078-356d0a0167ac\") " Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.179175 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60754672-3d3c-4763-8078-356d0a0167ac-kube-api-access-gvdhc" (OuterVolumeSpecName: "kube-api-access-gvdhc") pod "60754672-3d3c-4763-8078-356d0a0167ac" (UID: "60754672-3d3c-4763-8078-356d0a0167ac"). InnerVolumeSpecName "kube-api-access-gvdhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.273830 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvdhc\" (UniqueName: \"kubernetes.io/projected/60754672-3d3c-4763-8078-356d0a0167ac-kube-api-access-gvdhc\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.560936 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-b2fsv" event={"ID":"748546a6-1355-470f-b8d0-de395cf3f681","Type":"ContainerStarted","Data":"ee999697231a549c59c0cbcfc71d06fb72ac37908ba01d31423bef70468e1a7b"} Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.566269 4778 generic.go:334] "Generic (PLEG): container finished" podID="60754672-3d3c-4763-8078-356d0a0167ac" containerID="b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a" exitCode=0 Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.566389 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tbh2n" event={"ID":"60754672-3d3c-4763-8078-356d0a0167ac","Type":"ContainerDied","Data":"b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a"} Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.566428 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tbh2n" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.566498 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tbh2n" event={"ID":"60754672-3d3c-4763-8078-356d0a0167ac","Type":"ContainerDied","Data":"82b0ae343a0a10b9762a77419cf9048e88f45587940a8f410652a383e6167520"} Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.566520 4778 scope.go:117] "RemoveContainer" containerID="b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.590876 4778 scope.go:117] "RemoveContainer" containerID="b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.591009 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-b2fsv" podStartSLOduration=2.051891466 podStartE2EDuration="2.590974766s" podCreationTimestamp="2026-03-12 13:26:39 +0000 UTC" firstStartedPulling="2026-03-12 13:26:40.339046688 +0000 UTC m=+1018.787742084" lastFinishedPulling="2026-03-12 13:26:40.878129988 +0000 UTC m=+1019.326825384" observedRunningTime="2026-03-12 13:26:41.582094193 +0000 UTC m=+1020.030789599" watchObservedRunningTime="2026-03-12 13:26:41.590974766 +0000 UTC m=+1020.039670192" Mar 12 13:26:41 crc kubenswrapper[4778]: E0312 13:26:41.592048 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a\": container with ID starting with b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a not found: ID does not exist" containerID="b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.592118 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a"} err="failed to get container status \"b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a\": rpc error: code = NotFound desc = could not find container \"b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a\": container with ID starting with b26be7e9c9aca4fa6af7739dd64d7c833c232756970ba8b12fb55335ec5e133a not found: ID does not exist" Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.615048 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-tbh2n"] Mar 12 13:26:41 crc kubenswrapper[4778]: I0312 13:26:41.620746 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-tbh2n"] Mar 12 13:26:42 crc kubenswrapper[4778]: I0312 13:26:42.261711 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60754672-3d3c-4763-8078-356d0a0167ac" path="/var/lib/kubelet/pods/60754672-3d3c-4763-8078-356d0a0167ac/volumes" Mar 12 13:26:47 crc kubenswrapper[4778]: I0312 13:26:47.318505 4778 scope.go:117] "RemoveContainer" containerID="97b3a747ac158c0518500113b5af025bff04e06faaee081df03d1a06860f190f" Mar 12 13:26:49 crc kubenswrapper[4778]: I0312 13:26:49.886558 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:49 crc kubenswrapper[4778]: I0312 13:26:49.887038 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:49 crc kubenswrapper[4778]: I0312 13:26:49.937493 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:50 crc kubenswrapper[4778]: I0312 13:26:50.661301 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-b2fsv" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.793700 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr"] Mar 12 13:26:51 crc kubenswrapper[4778]: E0312 13:26:51.793968 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60754672-3d3c-4763-8078-356d0a0167ac" containerName="registry-server" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.793979 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="60754672-3d3c-4763-8078-356d0a0167ac" containerName="registry-server" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.794080 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="60754672-3d3c-4763-8078-356d0a0167ac" containerName="registry-server" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.795107 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.797143 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7k2vk" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.817751 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr"] Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.916964 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw2rg\" (UniqueName: \"kubernetes.io/projected/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-kube-api-access-bw2rg\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.917030 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-bundle\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:51 crc kubenswrapper[4778]: I0312 13:26:51.917110 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-util\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.018478 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw2rg\" (UniqueName: \"kubernetes.io/projected/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-kube-api-access-bw2rg\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.018593 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-bundle\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.018697 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-util\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.019632 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-bundle\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.019749 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-util\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.042215 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw2rg\" (UniqueName: \"kubernetes.io/projected/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-kube-api-access-bw2rg\") pod \"4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.114452 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.331277 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr"] Mar 12 13:26:52 crc kubenswrapper[4778]: I0312 13:26:52.641572 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" event={"ID":"e1d0ffee-229e-4da3-ac89-02bf6f6a439f","Type":"ContainerStarted","Data":"6c7b975bfe62559f47ed672e09903d56d09217ee6bc21f25b381ac2ab6bac8f1"} Mar 12 13:26:53 crc kubenswrapper[4778]: I0312 13:26:53.651704 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerID="f725f63e28c02b6398490e5b45fc89031bba591d75ca6d07df4522b25aa997b6" exitCode=0 Mar 12 13:26:53 crc kubenswrapper[4778]: I0312 13:26:53.651766 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" event={"ID":"e1d0ffee-229e-4da3-ac89-02bf6f6a439f","Type":"ContainerDied","Data":"f725f63e28c02b6398490e5b45fc89031bba591d75ca6d07df4522b25aa997b6"} Mar 12 13:26:55 crc kubenswrapper[4778]: I0312 13:26:55.668013 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerID="0fbe728cb5ced90554b79488aab511eb56de40f0946cf53e7cb505f322daef57" exitCode=0 Mar 12 13:26:55 crc kubenswrapper[4778]: I0312 13:26:55.668118 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" event={"ID":"e1d0ffee-229e-4da3-ac89-02bf6f6a439f","Type":"ContainerDied","Data":"0fbe728cb5ced90554b79488aab511eb56de40f0946cf53e7cb505f322daef57"} Mar 12 13:26:56 crc kubenswrapper[4778]: I0312 13:26:56.681090 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerID="55a5f5156e8222030ee1d0cca1ef9eee2f268a5a3accc79a2f3efdd69b8eb4d2" exitCode=0 Mar 12 13:26:56 crc kubenswrapper[4778]: I0312 13:26:56.681454 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" event={"ID":"e1d0ffee-229e-4da3-ac89-02bf6f6a439f","Type":"ContainerDied","Data":"55a5f5156e8222030ee1d0cca1ef9eee2f268a5a3accc79a2f3efdd69b8eb4d2"} Mar 12 13:26:57 crc kubenswrapper[4778]: I0312 13:26:57.924701 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.099226 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-bundle\") pod \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.099415 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw2rg\" (UniqueName: \"kubernetes.io/projected/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-kube-api-access-bw2rg\") pod \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.099482 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-util\") pod \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\" (UID: \"e1d0ffee-229e-4da3-ac89-02bf6f6a439f\") " Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.100280 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-bundle" (OuterVolumeSpecName: "bundle") pod "e1d0ffee-229e-4da3-ac89-02bf6f6a439f" (UID: "e1d0ffee-229e-4da3-ac89-02bf6f6a439f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.103502 4778 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.117667 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-kube-api-access-bw2rg" (OuterVolumeSpecName: "kube-api-access-bw2rg") pod "e1d0ffee-229e-4da3-ac89-02bf6f6a439f" (UID: "e1d0ffee-229e-4da3-ac89-02bf6f6a439f"). InnerVolumeSpecName "kube-api-access-bw2rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.209335 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw2rg\" (UniqueName: \"kubernetes.io/projected/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-kube-api-access-bw2rg\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.243871 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-util" (OuterVolumeSpecName: "util") pod "e1d0ffee-229e-4da3-ac89-02bf6f6a439f" (UID: "e1d0ffee-229e-4da3-ac89-02bf6f6a439f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.311167 4778 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d0ffee-229e-4da3-ac89-02bf6f6a439f-util\") on node \"crc\" DevicePath \"\"" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.696835 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" event={"ID":"e1d0ffee-229e-4da3-ac89-02bf6f6a439f","Type":"ContainerDied","Data":"6c7b975bfe62559f47ed672e09903d56d09217ee6bc21f25b381ac2ab6bac8f1"} Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.697350 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c7b975bfe62559f47ed672e09903d56d09217ee6bc21f25b381ac2ab6bac8f1" Mar 12 13:26:58 crc kubenswrapper[4778]: I0312 13:26:58.696923 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.002836 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl"] Mar 12 13:27:04 crc kubenswrapper[4778]: E0312 13:27:04.003263 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="extract" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.003274 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="extract" Mar 12 13:27:04 crc kubenswrapper[4778]: E0312 13:27:04.003285 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="util" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.003291 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="util" Mar 12 13:27:04 crc kubenswrapper[4778]: E0312 13:27:04.003304 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="pull" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.003310 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="pull" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.003407 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d0ffee-229e-4da3-ac89-02bf6f6a439f" containerName="extract" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.003785 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.006851 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-w82v7" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.039711 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl"] Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.196164 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klpjj\" (UniqueName: \"kubernetes.io/projected/34bbdc16-4518-4ee5-9a70-3cedcc5f0159-kube-api-access-klpjj\") pod \"openstack-operator-controller-init-5bc4df7446-x9bsl\" (UID: \"34bbdc16-4518-4ee5-9a70-3cedcc5f0159\") " pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.297648 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klpjj\" (UniqueName: \"kubernetes.io/projected/34bbdc16-4518-4ee5-9a70-3cedcc5f0159-kube-api-access-klpjj\") pod \"openstack-operator-controller-init-5bc4df7446-x9bsl\" (UID: \"34bbdc16-4518-4ee5-9a70-3cedcc5f0159\") " pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.329122 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klpjj\" (UniqueName: \"kubernetes.io/projected/34bbdc16-4518-4ee5-9a70-3cedcc5f0159-kube-api-access-klpjj\") pod \"openstack-operator-controller-init-5bc4df7446-x9bsl\" (UID: \"34bbdc16-4518-4ee5-9a70-3cedcc5f0159\") " pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:04 crc kubenswrapper[4778]: I0312 13:27:04.621510 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:05 crc kubenswrapper[4778]: I0312 13:27:05.037857 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl"] Mar 12 13:27:05 crc kubenswrapper[4778]: I0312 13:27:05.737923 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" event={"ID":"34bbdc16-4518-4ee5-9a70-3cedcc5f0159","Type":"ContainerStarted","Data":"b0cb81f3a436a1fa4c7d4d187d2d378e6115e3c52c453018d4a774d9e5554128"} Mar 12 13:27:12 crc kubenswrapper[4778]: I0312 13:27:12.873287 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" event={"ID":"34bbdc16-4518-4ee5-9a70-3cedcc5f0159","Type":"ContainerStarted","Data":"718457d5ecf5be484d18fcb2d14c3369f49cc01d76d7c6800297d67b914af438"} Mar 12 13:27:12 crc kubenswrapper[4778]: I0312 13:27:12.873826 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:24 crc kubenswrapper[4778]: I0312 13:27:24.624300 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" Mar 12 13:27:24 crc kubenswrapper[4778]: I0312 13:27:24.655996 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-5bc4df7446-x9bsl" podStartSLOduration=14.080995289 podStartE2EDuration="21.655977517s" podCreationTimestamp="2026-03-12 13:27:03 +0000 UTC" firstStartedPulling="2026-03-12 13:27:05.048986409 +0000 UTC m=+1043.497681795" lastFinishedPulling="2026-03-12 13:27:12.623968627 +0000 UTC m=+1051.072664023" observedRunningTime="2026-03-12 13:27:12.91531487 +0000 UTC m=+1051.364010276" watchObservedRunningTime="2026-03-12 13:27:24.655977517 +0000 UTC m=+1063.104672913" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.836767 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2"] Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.838848 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.846419 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2"] Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.865520 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wq5gn" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.886206 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc"] Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.887094 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.887838 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvw2p\" (UniqueName: \"kubernetes.io/projected/ffb8a1f4-4533-4368-a900-95d37fe1d3ad-kube-api-access-dvw2p\") pod \"barbican-operator-controller-manager-677bd678f7-6h2c2\" (UID: \"ffb8a1f4-4533-4368-a900-95d37fe1d3ad\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.887966 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljqr8\" (UniqueName: \"kubernetes.io/projected/c8818ac0-af8b-42c9-a923-425fe79ed203-kube-api-access-ljqr8\") pod \"cinder-operator-controller-manager-984cd4dcf-xm4cc\" (UID: \"c8818ac0-af8b-42c9-a923-425fe79ed203\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.895919 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-c2m28" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.992356 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvw2p\" (UniqueName: \"kubernetes.io/projected/ffb8a1f4-4533-4368-a900-95d37fe1d3ad-kube-api-access-dvw2p\") pod \"barbican-operator-controller-manager-677bd678f7-6h2c2\" (UID: \"ffb8a1f4-4533-4368-a900-95d37fe1d3ad\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:27:47 crc kubenswrapper[4778]: I0312 13:27:47.992471 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljqr8\" (UniqueName: \"kubernetes.io/projected/c8818ac0-af8b-42c9-a923-425fe79ed203-kube-api-access-ljqr8\") pod \"cinder-operator-controller-manager-984cd4dcf-xm4cc\" (UID: \"c8818ac0-af8b-42c9-a923-425fe79ed203\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.218482 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.219575 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.238566 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.257503 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qz45z" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.281352 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.291815 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.292957 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.317019 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvw2p\" (UniqueName: \"kubernetes.io/projected/ffb8a1f4-4533-4368-a900-95d37fe1d3ad-kube-api-access-dvw2p\") pod \"barbican-operator-controller-manager-677bd678f7-6h2c2\" (UID: \"ffb8a1f4-4533-4368-a900-95d37fe1d3ad\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.317048 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljqr8\" (UniqueName: \"kubernetes.io/projected/c8818ac0-af8b-42c9-a923-425fe79ed203-kube-api-access-ljqr8\") pod \"cinder-operator-controller-manager-984cd4dcf-xm4cc\" (UID: \"c8818ac0-af8b-42c9-a923-425fe79ed203\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.362587 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7zn8q" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.405071 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.419035 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7lr5\" (UniqueName: \"kubernetes.io/projected/ad531191-d7c5-4ef6-9929-3a5869751d98-kube-api-access-j7lr5\") pod \"designate-operator-controller-manager-66d56f6ff4-9n6jv\" (UID: \"ad531191-d7c5-4ef6-9929-3a5869751d98\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.419114 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hllx\" (UniqueName: \"kubernetes.io/projected/db7f6b97-2903-44bf-803f-c00c337400b9-kube-api-access-4hllx\") pod \"glance-operator-controller-manager-5964f64c48-gknp2\" (UID: \"db7f6b97-2903-44bf-803f-c00c337400b9\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.443533 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.444738 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.448621 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-dmq5h" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.469424 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.479253 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.480308 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.500470 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-54tdf" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.512733 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.523921 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.525793 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8jj6\" (UniqueName: \"kubernetes.io/projected/e290c1ea-a39d-451e-a24b-17a2b61ff6f0-kube-api-access-g8jj6\") pod \"heat-operator-controller-manager-77b6666d85-b7tkm\" (UID: \"e290c1ea-a39d-451e-a24b-17a2b61ff6f0\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.525841 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7lr5\" (UniqueName: \"kubernetes.io/projected/ad531191-d7c5-4ef6-9929-3a5869751d98-kube-api-access-j7lr5\") pod \"designate-operator-controller-manager-66d56f6ff4-9n6jv\" (UID: \"ad531191-d7c5-4ef6-9929-3a5869751d98\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.525873 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r2pb\" (UniqueName: \"kubernetes.io/projected/4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71-kube-api-access-4r2pb\") pod \"horizon-operator-controller-manager-6d9d6b584d-4jgt8\" (UID: \"4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.525916 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hllx\" (UniqueName: \"kubernetes.io/projected/db7f6b97-2903-44bf-803f-c00c337400b9-kube-api-access-4hllx\") pod \"glance-operator-controller-manager-5964f64c48-gknp2\" (UID: \"db7f6b97-2903-44bf-803f-c00c337400b9\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.558958 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.618747 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hllx\" (UniqueName: \"kubernetes.io/projected/db7f6b97-2903-44bf-803f-c00c337400b9-kube-api-access-4hllx\") pod \"glance-operator-controller-manager-5964f64c48-gknp2\" (UID: \"db7f6b97-2903-44bf-803f-c00c337400b9\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.621856 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7lr5\" (UniqueName: \"kubernetes.io/projected/ad531191-d7c5-4ef6-9929-3a5869751d98-kube-api-access-j7lr5\") pod \"designate-operator-controller-manager-66d56f6ff4-9n6jv\" (UID: \"ad531191-d7c5-4ef6-9929-3a5869751d98\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.627893 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz"] Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.629914 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.630303 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8jj6\" (UniqueName: \"kubernetes.io/projected/e290c1ea-a39d-451e-a24b-17a2b61ff6f0-kube-api-access-g8jj6\") pod \"heat-operator-controller-manager-77b6666d85-b7tkm\" (UID: \"e290c1ea-a39d-451e-a24b-17a2b61ff6f0\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.630497 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r2pb\" (UniqueName: \"kubernetes.io/projected/4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71-kube-api-access-4r2pb\") pod \"horizon-operator-controller-manager-6d9d6b584d-4jgt8\" (UID: \"4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.645510 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hb28j" Mar 12 13:27:48 crc kubenswrapper[4778]: I0312 13:27:48.650495 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.249043 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.258622 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.263472 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.265359 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpx49\" (UniqueName: \"kubernetes.io/projected/02bc06ca-f4e6-4fde-bd5d-882714d9652c-kube-api-access-vpx49\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.265437 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.279099 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8jj6\" (UniqueName: \"kubernetes.io/projected/e290c1ea-a39d-451e-a24b-17a2b61ff6f0-kube-api-access-g8jj6\") pod \"heat-operator-controller-manager-77b6666d85-b7tkm\" (UID: \"e290c1ea-a39d-451e-a24b-17a2b61ff6f0\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.372942 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpx49\" (UniqueName: \"kubernetes.io/projected/02bc06ca-f4e6-4fde-bd5d-882714d9652c-kube-api-access-vpx49\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.373373 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:49 crc kubenswrapper[4778]: E0312 13:27:49.373542 4778 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:49 crc kubenswrapper[4778]: E0312 13:27:49.373601 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert podName:02bc06ca-f4e6-4fde-bd5d-882714d9652c nodeName:}" failed. No retries permitted until 2026-03-12 13:27:49.873578983 +0000 UTC m=+1088.322274369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert") pod "infra-operator-controller-manager-5995f4446f-5d6qz" (UID: "02bc06ca-f4e6-4fde-bd5d-882714d9652c") : secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.373957 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.387972 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.399690 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpx49\" (UniqueName: \"kubernetes.io/projected/02bc06ca-f4e6-4fde-bd5d-882714d9652c-kube-api-access-vpx49\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.404328 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.410308 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.410623 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-gzpz9" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.418109 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.451120 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r2pb\" (UniqueName: \"kubernetes.io/projected/4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71-kube-api-access-4r2pb\") pod \"horizon-operator-controller-manager-6d9d6b584d-4jgt8\" (UID: \"4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.451448 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6hpjb" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.550595 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.717596 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.724351 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.725392 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.726329 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.726924 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.728434 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj8wf\" (UniqueName: \"kubernetes.io/projected/98a4cfbd-3037-48b5-9047-5d574dcc0aca-kube-api-access-mj8wf\") pod \"ironic-operator-controller-manager-6bbb499bbc-qb8s8\" (UID: \"98a4cfbd-3037-48b5-9047-5d574dcc0aca\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.728496 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bjmh\" (UniqueName: \"kubernetes.io/projected/7e02c37f-b9af-46c9-a743-03ead9b060db-kube-api-access-8bjmh\") pod \"keystone-operator-controller-manager-684f77d66d-7dxdh\" (UID: \"7e02c37f-b9af-46c9-a743-03ead9b060db\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.742647 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.743700 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.757505 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.757581 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.758766 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.767673 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-vbpwc" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.768127 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cgbz8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.768265 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-92tx4" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.796221 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.816160 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.818269 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.823118 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9nj6x" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.831384 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bjmh\" (UniqueName: \"kubernetes.io/projected/7e02c37f-b9af-46c9-a743-03ead9b060db-kube-api-access-8bjmh\") pod \"keystone-operator-controller-manager-684f77d66d-7dxdh\" (UID: \"7e02c37f-b9af-46c9-a743-03ead9b060db\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.831468 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb75x\" (UniqueName: \"kubernetes.io/projected/5e38a4fd-95f8-437b-923b-eca33b1387e6-kube-api-access-wb75x\") pod \"manila-operator-controller-manager-68f45f9d9f-pn8tk\" (UID: \"5e38a4fd-95f8-437b-923b-eca33b1387e6\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.831527 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks66w\" (UniqueName: \"kubernetes.io/projected/2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f-kube-api-access-ks66w\") pod \"mariadb-operator-controller-manager-658d4cdd5-jlbft\" (UID: \"2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.831572 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw6mz\" (UniqueName: \"kubernetes.io/projected/d7288cc6-4247-4d03-bd37-9862243bf613-kube-api-access-qw6mz\") pod \"nova-operator-controller-manager-686d5f9fbd-vv9rc\" (UID: \"d7288cc6-4247-4d03-bd37-9862243bf613\") " pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.831598 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txndm\" (UniqueName: \"kubernetes.io/projected/076835c9-352b-4e40-80c4-3bce3bb80594-kube-api-access-txndm\") pod \"neutron-operator-controller-manager-776c5696bf-dd2ft\" (UID: \"076835c9-352b-4e40-80c4-3bce3bb80594\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.831644 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj8wf\" (UniqueName: \"kubernetes.io/projected/98a4cfbd-3037-48b5-9047-5d574dcc0aca-kube-api-access-mj8wf\") pod \"ironic-operator-controller-manager-6bbb499bbc-qb8s8\" (UID: \"98a4cfbd-3037-48b5-9047-5d574dcc0aca\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.854366 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.856683 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj8wf\" (UniqueName: \"kubernetes.io/projected/98a4cfbd-3037-48b5-9047-5d574dcc0aca-kube-api-access-mj8wf\") pod \"ironic-operator-controller-manager-6bbb499bbc-qb8s8\" (UID: \"98a4cfbd-3037-48b5-9047-5d574dcc0aca\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.872478 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.873297 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.881374 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-j5jx9" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.885598 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bjmh\" (UniqueName: \"kubernetes.io/projected/7e02c37f-b9af-46c9-a743-03ead9b060db-kube-api-access-8bjmh\") pod \"keystone-operator-controller-manager-684f77d66d-7dxdh\" (UID: \"7e02c37f-b9af-46c9-a743-03ead9b060db\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.907689 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.933569 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks66w\" (UniqueName: \"kubernetes.io/projected/2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f-kube-api-access-ks66w\") pod \"mariadb-operator-controller-manager-658d4cdd5-jlbft\" (UID: \"2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.933626 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw6mz\" (UniqueName: \"kubernetes.io/projected/d7288cc6-4247-4d03-bd37-9862243bf613-kube-api-access-qw6mz\") pod \"nova-operator-controller-manager-686d5f9fbd-vv9rc\" (UID: \"d7288cc6-4247-4d03-bd37-9862243bf613\") " pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.933647 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txndm\" (UniqueName: \"kubernetes.io/projected/076835c9-352b-4e40-80c4-3bce3bb80594-kube-api-access-txndm\") pod \"neutron-operator-controller-manager-776c5696bf-dd2ft\" (UID: \"076835c9-352b-4e40-80c4-3bce3bb80594\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.933709 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.933741 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb75x\" (UniqueName: \"kubernetes.io/projected/5e38a4fd-95f8-437b-923b-eca33b1387e6-kube-api-access-wb75x\") pod \"manila-operator-controller-manager-68f45f9d9f-pn8tk\" (UID: \"5e38a4fd-95f8-437b-923b-eca33b1387e6\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.933768 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sl2m\" (UniqueName: \"kubernetes.io/projected/1a01d06c-be6f-45de-a22d-c8f1058a3a84-kube-api-access-5sl2m\") pod \"octavia-operator-controller-manager-5f4f55cb5c-cdgg9\" (UID: \"1a01d06c-be6f-45de-a22d-c8f1058a3a84\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:27:49 crc kubenswrapper[4778]: E0312 13:27:49.934237 4778 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:49 crc kubenswrapper[4778]: E0312 13:27:49.934278 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert podName:02bc06ca-f4e6-4fde-bd5d-882714d9652c nodeName:}" failed. No retries permitted until 2026-03-12 13:27:50.934262808 +0000 UTC m=+1089.382958204 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert") pod "infra-operator-controller-manager-5995f4446f-5d6qz" (UID: "02bc06ca-f4e6-4fde-bd5d-882714d9652c") : secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.955746 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw6mz\" (UniqueName: \"kubernetes.io/projected/d7288cc6-4247-4d03-bd37-9862243bf613-kube-api-access-qw6mz\") pod \"nova-operator-controller-manager-686d5f9fbd-vv9rc\" (UID: \"d7288cc6-4247-4d03-bd37-9862243bf613\") " pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.971429 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.972515 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.976561 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks66w\" (UniqueName: \"kubernetes.io/projected/2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f-kube-api-access-ks66w\") pod \"mariadb-operator-controller-manager-658d4cdd5-jlbft\" (UID: \"2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.976868 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-mv9px" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.980464 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txndm\" (UniqueName: \"kubernetes.io/projected/076835c9-352b-4e40-80c4-3bce3bb80594-kube-api-access-txndm\") pod \"neutron-operator-controller-manager-776c5696bf-dd2ft\" (UID: \"076835c9-352b-4e40-80c4-3bce3bb80594\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.984923 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.987632 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6"] Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.991955 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb75x\" (UniqueName: \"kubernetes.io/projected/5e38a4fd-95f8-437b-923b-eca33b1387e6-kube-api-access-wb75x\") pod \"manila-operator-controller-manager-68f45f9d9f-pn8tk\" (UID: \"5e38a4fd-95f8-437b-923b-eca33b1387e6\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:27:49 crc kubenswrapper[4778]: I0312 13:27:49.995691 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.006760 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lmw9q" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.006933 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.016357 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.021869 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.023100 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.025474 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.025615 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-m7kt4" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.026148 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.031833 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-84mps"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.033614 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.038221 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sl2m\" (UniqueName: \"kubernetes.io/projected/1a01d06c-be6f-45de-a22d-c8f1058a3a84-kube-api-access-5sl2m\") pod \"octavia-operator-controller-manager-5f4f55cb5c-cdgg9\" (UID: \"1a01d06c-be6f-45de-a22d-c8f1058a3a84\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.041134 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fdhvv" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.119853 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.277252 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-764zg\" (UniqueName: \"kubernetes.io/projected/52524252-25bd-49e5-822e-3d4668aff2f9-kube-api-access-764zg\") pod \"placement-operator-controller-manager-574d45c66c-wvpf8\" (UID: \"52524252-25bd-49e5-822e-3d4668aff2f9\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.278501 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.278993 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.278358 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sl2m\" (UniqueName: \"kubernetes.io/projected/1a01d06c-be6f-45de-a22d-c8f1058a3a84-kube-api-access-5sl2m\") pod \"octavia-operator-controller-manager-5f4f55cb5c-cdgg9\" (UID: \"1a01d06c-be6f-45de-a22d-c8f1058a3a84\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.279618 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.284317 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n48j5\" (UniqueName: \"kubernetes.io/projected/4f7d316e-6896-4f84-8423-6f79778c1c6b-kube-api-access-n48j5\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.333029 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktk6s\" (UniqueName: \"kubernetes.io/projected/8d38fd7e-6fa1-4b0c-9c82-9c57290c7837-kube-api-access-ktk6s\") pod \"ovn-operator-controller-manager-bbc5b68f9-bbgmb\" (UID: \"8d38fd7e-6fa1-4b0c-9c82-9c57290c7837\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.333131 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-446sv\" (UniqueName: \"kubernetes.io/projected/64a36384-f2e6-4077-b2ca-de2a6ce6ea06-kube-api-access-446sv\") pod \"swift-operator-controller-manager-677c674df7-84mps\" (UID: \"64a36384-f2e6-4077-b2ca-de2a6ce6ea06\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.333169 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.373501 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.754812 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.755127 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.755833 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.762512 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-84mps"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.772920 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktk6s\" (UniqueName: \"kubernetes.io/projected/8d38fd7e-6fa1-4b0c-9c82-9c57290c7837-kube-api-access-ktk6s\") pod \"ovn-operator-controller-manager-bbc5b68f9-bbgmb\" (UID: \"8d38fd7e-6fa1-4b0c-9c82-9c57290c7837\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.772954 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-446sv\" (UniqueName: \"kubernetes.io/projected/64a36384-f2e6-4077-b2ca-de2a6ce6ea06-kube-api-access-446sv\") pod \"swift-operator-controller-manager-677c674df7-84mps\" (UID: \"64a36384-f2e6-4077-b2ca-de2a6ce6ea06\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.772986 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.773029 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-764zg\" (UniqueName: \"kubernetes.io/projected/52524252-25bd-49e5-822e-3d4668aff2f9-kube-api-access-764zg\") pod \"placement-operator-controller-manager-574d45c66c-wvpf8\" (UID: \"52524252-25bd-49e5-822e-3d4668aff2f9\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.773059 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n48j5\" (UniqueName: \"kubernetes.io/projected/4f7d316e-6896-4f84-8423-6f79778c1c6b-kube-api-access-n48j5\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:50 crc kubenswrapper[4778]: E0312 13:27:50.773356 4778 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:50 crc kubenswrapper[4778]: E0312 13:27:50.773399 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert podName:4f7d316e-6896-4f84-8423-6f79778c1c6b nodeName:}" failed. No retries permitted until 2026-03-12 13:27:51.273386589 +0000 UTC m=+1089.722081975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" (UID: "4f7d316e-6896-4f84-8423-6f79778c1c6b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.784356 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-j8lzm" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.784653 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.793950 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.795474 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.795622 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.804532 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6fvqh" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.824220 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.825826 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-764zg\" (UniqueName: \"kubernetes.io/projected/52524252-25bd-49e5-822e-3d4668aff2f9-kube-api-access-764zg\") pod \"placement-operator-controller-manager-574d45c66c-wvpf8\" (UID: \"52524252-25bd-49e5-822e-3d4668aff2f9\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.826782 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-446sv\" (UniqueName: \"kubernetes.io/projected/64a36384-f2e6-4077-b2ca-de2a6ce6ea06-kube-api-access-446sv\") pod \"swift-operator-controller-manager-677c674df7-84mps\" (UID: \"64a36384-f2e6-4077-b2ca-de2a6ce6ea06\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.839200 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n48j5\" (UniqueName: \"kubernetes.io/projected/4f7d316e-6896-4f84-8423-6f79778c1c6b-kube-api-access-n48j5\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.844947 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktk6s\" (UniqueName: \"kubernetes.io/projected/8d38fd7e-6fa1-4b0c-9c82-9c57290c7837-kube-api-access-ktk6s\") pod \"ovn-operator-controller-manager-bbc5b68f9-bbgmb\" (UID: \"8d38fd7e-6fa1-4b0c-9c82-9c57290c7837\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.854609 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.854696 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.859644 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-pvdlr" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.873857 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.874204 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.875667 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ql6\" (UniqueName: \"kubernetes.io/projected/6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c-kube-api-access-d2ql6\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-gfv5z\" (UID: \"6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.889130 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.889891 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.897559 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.915757 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc"] Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.938815 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-j8vfd" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.938885 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.939002 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978384 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978443 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ql6\" (UniqueName: \"kubernetes.io/projected/6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c-kube-api-access-d2ql6\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-gfv5z\" (UID: \"6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978463 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978493 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7fsj\" (UniqueName: \"kubernetes.io/projected/8c02ecb8-0e15-4672-823a-c4437ca5bf8c-kube-api-access-j7fsj\") pod \"watcher-operator-controller-manager-6dd88c6f67-2tjsk\" (UID: \"8c02ecb8-0e15-4672-823a-c4437ca5bf8c\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978520 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gz2m\" (UniqueName: \"kubernetes.io/projected/ed9b9271-4ae9-440a-9411-15d46267106e-kube-api-access-9gz2m\") pod \"test-operator-controller-manager-5c5cb9c4d7-pcfrz\" (UID: \"ed9b9271-4ae9-440a-9411-15d46267106e\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978537 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:27:50 crc kubenswrapper[4778]: E0312 13:27:50.978655 4778 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:50 crc kubenswrapper[4778]: E0312 13:27:50.978697 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert podName:02bc06ca-f4e6-4fde-bd5d-882714d9652c nodeName:}" failed. No retries permitted until 2026-03-12 13:27:52.978683117 +0000 UTC m=+1091.427378513 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert") pod "infra-operator-controller-manager-5995f4446f-5d6qz" (UID: "02bc06ca-f4e6-4fde-bd5d-882714d9652c") : secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.978550 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:50 crc kubenswrapper[4778]: I0312 13:27:50.979243 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbp8p\" (UniqueName: \"kubernetes.io/projected/d0784623-5f08-4109-9c7e-0a329210ce07-kube-api-access-vbp8p\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.010879 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ql6\" (UniqueName: \"kubernetes.io/projected/6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c-kube-api-access-d2ql6\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-gfv5z\" (UID: \"6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.024359 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm"] Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.068098 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b"] Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.069065 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.072261 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-625vs" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.072595 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.080696 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gz2m\" (UniqueName: \"kubernetes.io/projected/ed9b9271-4ae9-440a-9411-15d46267106e-kube-api-access-9gz2m\") pod \"test-operator-controller-manager-5c5cb9c4d7-pcfrz\" (UID: \"ed9b9271-4ae9-440a-9411-15d46267106e\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.080764 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbp8p\" (UniqueName: \"kubernetes.io/projected/d0784623-5f08-4109-9c7e-0a329210ce07-kube-api-access-vbp8p\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.080809 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.080846 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.080875 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7fsj\" (UniqueName: \"kubernetes.io/projected/8c02ecb8-0e15-4672-823a-c4437ca5bf8c-kube-api-access-j7fsj\") pod \"watcher-operator-controller-manager-6dd88c6f67-2tjsk\" (UID: \"8c02ecb8-0e15-4672-823a-c4437ca5bf8c\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.081140 4778 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.081177 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:51.58116298 +0000 UTC m=+1090.029858376 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "webhook-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.081279 4778 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.081301 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:51.581295134 +0000 UTC m=+1090.029990530 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "metrics-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.089292 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.094732 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b"] Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.098710 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7fsj\" (UniqueName: \"kubernetes.io/projected/8c02ecb8-0e15-4672-823a-c4437ca5bf8c-kube-api-access-j7fsj\") pod \"watcher-operator-controller-manager-6dd88c6f67-2tjsk\" (UID: \"8c02ecb8-0e15-4672-823a-c4437ca5bf8c\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.112334 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbp8p\" (UniqueName: \"kubernetes.io/projected/d0784623-5f08-4109-9c7e-0a329210ce07-kube-api-access-vbp8p\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.115205 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gz2m\" (UniqueName: \"kubernetes.io/projected/ed9b9271-4ae9-440a-9411-15d46267106e-kube-api-access-9gz2m\") pod \"test-operator-controller-manager-5c5cb9c4d7-pcfrz\" (UID: \"ed9b9271-4ae9-440a-9411-15d46267106e\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:27:51 crc kubenswrapper[4778]: W0312 13:27:51.153686 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode290c1ea_a39d_451e_a24b_17a2b61ff6f0.slice/crio-29b3719bed2f4fb41361bea063a0c860366f43adee3b02d65b29b84f9aa6079b WatchSource:0}: Error finding container 29b3719bed2f4fb41361bea063a0c860366f43adee3b02d65b29b84f9aa6079b: Status 404 returned error can't find the container with id 29b3719bed2f4fb41361bea063a0c860366f43adee3b02d65b29b84f9aa6079b Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.533271 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.533742 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.552339 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.552467 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldr22\" (UniqueName: \"kubernetes.io/projected/034f39d8-a33e-4e37-bcde-51fb22debdd1-kube-api-access-ldr22\") pod \"rabbitmq-cluster-operator-manager-668c99d594-shf7b\" (UID: \"034f39d8-a33e-4e37-bcde-51fb22debdd1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.552677 4778 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.552732 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert podName:4f7d316e-6896-4f84-8423-6f79778c1c6b nodeName:}" failed. No retries permitted until 2026-03-12 13:27:52.552714644 +0000 UTC m=+1091.001410030 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" (UID: "4f7d316e-6896-4f84-8423-6f79778c1c6b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.637850 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8"] Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.659718 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.660152 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldr22\" (UniqueName: \"kubernetes.io/projected/034f39d8-a33e-4e37-bcde-51fb22debdd1-kube-api-access-ldr22\") pod \"rabbitmq-cluster-operator-manager-668c99d594-shf7b\" (UID: \"034f39d8-a33e-4e37-bcde-51fb22debdd1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.660298 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.659860 4778 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.661241 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:52.660817118 +0000 UTC m=+1091.109512514 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "metrics-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.661393 4778 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: E0312 13:27:51.661423 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:52.661412925 +0000 UTC m=+1091.110108311 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "webhook-server-cert" not found Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.834136 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldr22\" (UniqueName: \"kubernetes.io/projected/034f39d8-a33e-4e37-bcde-51fb22debdd1-kube-api-access-ldr22\") pod \"rabbitmq-cluster-operator-manager-668c99d594-shf7b\" (UID: \"034f39d8-a33e-4e37-bcde-51fb22debdd1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.858255 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" event={"ID":"e290c1ea-a39d-451e-a24b-17a2b61ff6f0","Type":"ContainerStarted","Data":"29b3719bed2f4fb41361bea063a0c860366f43adee3b02d65b29b84f9aa6079b"} Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.861197 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" event={"ID":"ffb8a1f4-4533-4368-a900-95d37fe1d3ad","Type":"ContainerStarted","Data":"6add4468a8549d66173c2ee7bf0009205051ad85b83c631387eac0d7eac012be"} Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.863319 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" event={"ID":"c8818ac0-af8b-42c9-a923-425fe79ed203","Type":"ContainerStarted","Data":"dbd1b8e4778d42734e0b1ca1d10ae31267885adbdbe5b26c477fe6eac22e7012"} Mar 12 13:27:51 crc kubenswrapper[4778]: I0312 13:27:51.886879 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2"] Mar 12 13:27:51 crc kubenswrapper[4778]: W0312 13:27:51.891936 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c2bf703_ecc1_4bb1_aa03_a64e55dfdb71.slice/crio-afc92afd611cd768f64e952660cb89d76a0eb88173655f02b34c045043ee1d51 WatchSource:0}: Error finding container afc92afd611cd768f64e952660cb89d76a0eb88173655f02b34c045043ee1d51: Status 404 returned error can't find the container with id afc92afd611cd768f64e952660cb89d76a0eb88173655f02b34c045043ee1d51 Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.274870 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.554778 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:52 crc kubenswrapper[4778]: E0312 13:27:52.555013 4778 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:52 crc kubenswrapper[4778]: E0312 13:27:52.555060 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert podName:4f7d316e-6896-4f84-8423-6f79778c1c6b nodeName:}" failed. No retries permitted until 2026-03-12 13:27:54.555046181 +0000 UTC m=+1093.003741577 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" (UID: "4f7d316e-6896-4f84-8423-6f79778c1c6b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.669434 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.669783 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:52 crc kubenswrapper[4778]: E0312 13:27:52.670124 4778 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 13:27:52 crc kubenswrapper[4778]: E0312 13:27:52.670226 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:54.670204217 +0000 UTC m=+1093.118899613 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "webhook-server-cert" not found Mar 12 13:27:52 crc kubenswrapper[4778]: E0312 13:27:52.670820 4778 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 13:27:52 crc kubenswrapper[4778]: E0312 13:27:52.670865 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:54.670850555 +0000 UTC m=+1093.119545951 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "metrics-server-cert" not found Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.689129 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv"] Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.988321 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" event={"ID":"ad531191-d7c5-4ef6-9929-3a5869751d98","Type":"ContainerStarted","Data":"03fd176a7780ae59f406e8483eca6a0c49fd1aa0b68aa0a58f0ac7631dbdd75a"} Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.989828 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" event={"ID":"db7f6b97-2903-44bf-803f-c00c337400b9","Type":"ContainerStarted","Data":"462202b2866624ca0ac8bfe0fcd0546313eca0b1abcde399dfb254602bc2653a"} Mar 12 13:27:52 crc kubenswrapper[4778]: I0312 13:27:52.992337 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" event={"ID":"4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71","Type":"ContainerStarted","Data":"afc92afd611cd768f64e952660cb89d76a0eb88173655f02b34c045043ee1d51"} Mar 12 13:27:53 crc kubenswrapper[4778]: I0312 13:27:53.022768 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:53 crc kubenswrapper[4778]: E0312 13:27:53.022941 4778 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:53 crc kubenswrapper[4778]: E0312 13:27:53.022985 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert podName:02bc06ca-f4e6-4fde-bd5d-882714d9652c nodeName:}" failed. No retries permitted until 2026-03-12 13:27:57.022971352 +0000 UTC m=+1095.471666748 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert") pod "infra-operator-controller-manager-5995f4446f-5d6qz" (UID: "02bc06ca-f4e6-4fde-bd5d-882714d9652c") : secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:53 crc kubenswrapper[4778]: I0312 13:27:53.103050 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8"] Mar 12 13:27:53 crc kubenswrapper[4778]: W0312 13:27:53.334396 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98a4cfbd_3037_48b5_9047_5d574dcc0aca.slice/crio-d86fe0c7300172809b88b6373a2c044ec040f3664871890f725be0ab41c29b5d WatchSource:0}: Error finding container d86fe0c7300172809b88b6373a2c044ec040f3664871890f725be0ab41c29b5d: Status 404 returned error can't find the container with id d86fe0c7300172809b88b6373a2c044ec040f3664871890f725be0ab41c29b5d Mar 12 13:27:53 crc kubenswrapper[4778]: I0312 13:27:53.796030 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc"] Mar 12 13:27:53 crc kubenswrapper[4778]: I0312 13:27:53.976292 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh"] Mar 12 13:27:53 crc kubenswrapper[4778]: I0312 13:27:53.990466 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk"] Mar 12 13:27:53 crc kubenswrapper[4778]: I0312 13:27:53.996693 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft"] Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.035937 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" event={"ID":"7e02c37f-b9af-46c9-a743-03ead9b060db","Type":"ContainerStarted","Data":"bb467c1fe68b7f46d0ab121ad5687ab6f4dfc931fbf41a29126eda47416b8c3b"} Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.067431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" event={"ID":"98a4cfbd-3037-48b5-9047-5d574dcc0aca","Type":"ContainerStarted","Data":"d86fe0c7300172809b88b6373a2c044ec040f3664871890f725be0ab41c29b5d"} Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.525660 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9"] Mar 12 13:27:54 crc kubenswrapper[4778]: W0312 13:27:54.545480 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a01d06c_be6f_45de_a22d_c8f1058a3a84.slice/crio-893de7cce2036fa6907343b24e58e80f6d7c444f81fe16929220418fbe2057d2 WatchSource:0}: Error finding container 893de7cce2036fa6907343b24e58e80f6d7c444f81fe16929220418fbe2057d2: Status 404 returned error can't find the container with id 893de7cce2036fa6907343b24e58e80f6d7c444f81fe16929220418fbe2057d2 Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.705982 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.706027 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.706066 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:54 crc kubenswrapper[4778]: E0312 13:27:54.706200 4778 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 13:27:54 crc kubenswrapper[4778]: E0312 13:27:54.706594 4778 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:54 crc kubenswrapper[4778]: E0312 13:27:54.706634 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert podName:4f7d316e-6896-4f84-8423-6f79778c1c6b nodeName:}" failed. No retries permitted until 2026-03-12 13:27:58.706620866 +0000 UTC m=+1097.155316262 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" (UID: "4f7d316e-6896-4f84-8423-6f79778c1c6b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:54 crc kubenswrapper[4778]: E0312 13:27:54.706905 4778 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 13:27:54 crc kubenswrapper[4778]: E0312 13:27:54.706930 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:58.706922515 +0000 UTC m=+1097.155617911 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "webhook-server-cert" not found Mar 12 13:27:54 crc kubenswrapper[4778]: E0312 13:27:54.706952 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:27:58.706945356 +0000 UTC m=+1097.155640752 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "metrics-server-cert" not found Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.802900 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8"] Mar 12 13:27:54 crc kubenswrapper[4778]: W0312 13:27:54.822957 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52524252_25bd_49e5_822e_3d4668aff2f9.slice/crio-a3262b7c97c85db32c9ca2463df50a7754cb0cf38f9f4881e25a720c68d6db8e WatchSource:0}: Error finding container a3262b7c97c85db32c9ca2463df50a7754cb0cf38f9f4881e25a720c68d6db8e: Status 404 returned error can't find the container with id a3262b7c97c85db32c9ca2463df50a7754cb0cf38f9f4881e25a720c68d6db8e Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.835451 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb"] Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.858776 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft"] Mar 12 13:27:54 crc kubenswrapper[4778]: I0312 13:27:54.877085 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z"] Mar 12 13:27:54 crc kubenswrapper[4778]: W0312 13:27:54.885309 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d577800_0ee1_4fe5_a7fb_8794fb8c4c6f.slice/crio-1f781822864f6ccaee041de62ccaae7e12f9cfdd25380060e8173a78f59408eb WatchSource:0}: Error finding container 1f781822864f6ccaee041de62ccaae7e12f9cfdd25380060e8173a78f59408eb: Status 404 returned error can't find the container with id 1f781822864f6ccaee041de62ccaae7e12f9cfdd25380060e8173a78f59408eb Mar 12 13:27:54 crc kubenswrapper[4778]: W0312 13:27:54.893361 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ad9bf9f_7214_44bc_a65d_1dcbf385fc2c.slice/crio-66a35a20b3f89741667ee8b69c74ea7489e9678aad111e22ae5c6b16591bf61c WatchSource:0}: Error finding container 66a35a20b3f89741667ee8b69c74ea7489e9678aad111e22ae5c6b16591bf61c: Status 404 returned error can't find the container with id 66a35a20b3f89741667ee8b69c74ea7489e9678aad111e22ae5c6b16591bf61c Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.189152 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" event={"ID":"d7288cc6-4247-4d03-bd37-9862243bf613","Type":"ContainerStarted","Data":"15b7f4fe8a5cfc1726f6b38dfbf4ec237beccfbdf2b5296e977aa567cf3ea730"} Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.210321 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" event={"ID":"52524252-25bd-49e5-822e-3d4668aff2f9","Type":"ContainerStarted","Data":"a3262b7c97c85db32c9ca2463df50a7754cb0cf38f9f4881e25a720c68d6db8e"} Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.214713 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz"] Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.215124 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" event={"ID":"1a01d06c-be6f-45de-a22d-c8f1058a3a84","Type":"ContainerStarted","Data":"893de7cce2036fa6907343b24e58e80f6d7c444f81fe16929220418fbe2057d2"} Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.220687 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" event={"ID":"076835c9-352b-4e40-80c4-3bce3bb80594","Type":"ContainerStarted","Data":"4c641f62d7c300472a2c0a75691207f30010390a30bbf20056fede889e396c80"} Mar 12 13:27:55 crc kubenswrapper[4778]: W0312 13:27:55.240755 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded9b9271_4ae9_440a_9411_15d46267106e.slice/crio-34158edfb6803522b4b6384fa1afdafb1dff9bfbaeeaa1e79de5c066d21cc548 WatchSource:0}: Error finding container 34158edfb6803522b4b6384fa1afdafb1dff9bfbaeeaa1e79de5c066d21cc548: Status 404 returned error can't find the container with id 34158edfb6803522b4b6384fa1afdafb1dff9bfbaeeaa1e79de5c066d21cc548 Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.244995 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" event={"ID":"5e38a4fd-95f8-437b-923b-eca33b1387e6","Type":"ContainerStarted","Data":"57c55a7e15d470c6379374ab81f4e9c300b6344a7c41c1839187c215c09c56a1"} Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.251911 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk"] Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.253715 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" event={"ID":"8d38fd7e-6fa1-4b0c-9c82-9c57290c7837","Type":"ContainerStarted","Data":"f9ab4ff66ea9e47b1e89bf37ec2b3235d3ef42604bc55019cfd4909c095e9464"} Mar 12 13:27:55 crc kubenswrapper[4778]: W0312 13:27:55.257636 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c02ecb8_0e15_4672_823a_c4437ca5bf8c.slice/crio-706f249840e7b1d704766e2b3527abb545d3f6ae1d202df51e694140eef30e9b WatchSource:0}: Error finding container 706f249840e7b1d704766e2b3527abb545d3f6ae1d202df51e694140eef30e9b: Status 404 returned error can't find the container with id 706f249840e7b1d704766e2b3527abb545d3f6ae1d202df51e694140eef30e9b Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.260576 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b"] Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.268864 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" event={"ID":"2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f","Type":"ContainerStarted","Data":"1f781822864f6ccaee041de62ccaae7e12f9cfdd25380060e8173a78f59408eb"} Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.270562 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-84mps"] Mar 12 13:27:55 crc kubenswrapper[4778]: I0312 13:27:55.273962 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" event={"ID":"6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c","Type":"ContainerStarted","Data":"66a35a20b3f89741667ee8b69c74ea7489e9678aad111e22ae5c6b16591bf61c"} Mar 12 13:27:55 crc kubenswrapper[4778]: W0312 13:27:55.316913 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64a36384_f2e6_4077_b2ca_de2a6ce6ea06.slice/crio-93919bb9cce297fa177b1164e22870e7a66bf860567fe714205d8702b80bd9a2 WatchSource:0}: Error finding container 93919bb9cce297fa177b1164e22870e7a66bf860567fe714205d8702b80bd9a2: Status 404 returned error can't find the container with id 93919bb9cce297fa177b1164e22870e7a66bf860567fe714205d8702b80bd9a2 Mar 12 13:27:55 crc kubenswrapper[4778]: W0312 13:27:55.338795 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod034f39d8_a33e_4e37_bcde_51fb22debdd1.slice/crio-7dba3190ef6cfd18ad725f320f6050312ce06ba17d117d060eed0d71ed1adc55 WatchSource:0}: Error finding container 7dba3190ef6cfd18ad725f320f6050312ce06ba17d117d060eed0d71ed1adc55: Status 404 returned error can't find the container with id 7dba3190ef6cfd18ad725f320f6050312ce06ba17d117d060eed0d71ed1adc55 Mar 12 13:27:56 crc kubenswrapper[4778]: I0312 13:27:56.388666 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" event={"ID":"034f39d8-a33e-4e37-bcde-51fb22debdd1","Type":"ContainerStarted","Data":"7dba3190ef6cfd18ad725f320f6050312ce06ba17d117d060eed0d71ed1adc55"} Mar 12 13:27:56 crc kubenswrapper[4778]: I0312 13:27:56.417426 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" event={"ID":"8c02ecb8-0e15-4672-823a-c4437ca5bf8c","Type":"ContainerStarted","Data":"706f249840e7b1d704766e2b3527abb545d3f6ae1d202df51e694140eef30e9b"} Mar 12 13:27:56 crc kubenswrapper[4778]: I0312 13:27:56.439107 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" event={"ID":"ed9b9271-4ae9-440a-9411-15d46267106e","Type":"ContainerStarted","Data":"34158edfb6803522b4b6384fa1afdafb1dff9bfbaeeaa1e79de5c066d21cc548"} Mar 12 13:27:56 crc kubenswrapper[4778]: I0312 13:27:56.440534 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" event={"ID":"64a36384-f2e6-4077-b2ca-de2a6ce6ea06","Type":"ContainerStarted","Data":"93919bb9cce297fa177b1164e22870e7a66bf860567fe714205d8702b80bd9a2"} Mar 12 13:27:57 crc kubenswrapper[4778]: I0312 13:27:57.084816 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:27:57 crc kubenswrapper[4778]: E0312 13:27:57.085054 4778 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:57 crc kubenswrapper[4778]: E0312 13:27:57.085114 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert podName:02bc06ca-f4e6-4fde-bd5d-882714d9652c nodeName:}" failed. No retries permitted until 2026-03-12 13:28:05.085096695 +0000 UTC m=+1103.533792091 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert") pod "infra-operator-controller-manager-5995f4446f-5d6qz" (UID: "02bc06ca-f4e6-4fde-bd5d-882714d9652c") : secret "infra-operator-webhook-server-cert" not found Mar 12 13:27:59 crc kubenswrapper[4778]: I0312 13:27:59.314409 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:59 crc kubenswrapper[4778]: I0312 13:27:59.315067 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:27:59 crc kubenswrapper[4778]: I0312 13:27:59.315121 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:27:59 crc kubenswrapper[4778]: E0312 13:27:59.343103 4778 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:59 crc kubenswrapper[4778]: E0312 13:27:59.343170 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert podName:4f7d316e-6896-4f84-8423-6f79778c1c6b nodeName:}" failed. No retries permitted until 2026-03-12 13:28:07.343153349 +0000 UTC m=+1105.791848745 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" (UID: "4f7d316e-6896-4f84-8423-6f79778c1c6b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 13:27:59 crc kubenswrapper[4778]: E0312 13:27:59.353177 4778 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 13:27:59 crc kubenswrapper[4778]: E0312 13:27:59.353470 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:28:07.353445952 +0000 UTC m=+1105.802141348 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "metrics-server-cert" not found Mar 12 13:27:59 crc kubenswrapper[4778]: E0312 13:27:59.353200 4778 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 13:27:59 crc kubenswrapper[4778]: E0312 13:27:59.353700 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:28:07.353686899 +0000 UTC m=+1105.802382295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "webhook-server-cert" not found Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.126907 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555368-d2cpg"] Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.127858 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.129998 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.130222 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.130986 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.143914 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555368-d2cpg"] Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.225595 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m2kn\" (UniqueName: \"kubernetes.io/projected/20d587ee-7b57-4b99-a800-c6d46322d799-kube-api-access-9m2kn\") pod \"auto-csr-approver-29555368-d2cpg\" (UID: \"20d587ee-7b57-4b99-a800-c6d46322d799\") " pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.338303 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m2kn\" (UniqueName: \"kubernetes.io/projected/20d587ee-7b57-4b99-a800-c6d46322d799-kube-api-access-9m2kn\") pod \"auto-csr-approver-29555368-d2cpg\" (UID: \"20d587ee-7b57-4b99-a800-c6d46322d799\") " pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.371695 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m2kn\" (UniqueName: \"kubernetes.io/projected/20d587ee-7b57-4b99-a800-c6d46322d799-kube-api-access-9m2kn\") pod \"auto-csr-approver-29555368-d2cpg\" (UID: \"20d587ee-7b57-4b99-a800-c6d46322d799\") " pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:00 crc kubenswrapper[4778]: I0312 13:28:00.530162 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:01 crc kubenswrapper[4778]: I0312 13:28:01.680903 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555368-d2cpg"] Mar 12 13:28:01 crc kubenswrapper[4778]: I0312 13:28:01.855517 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" event={"ID":"20d587ee-7b57-4b99-a800-c6d46322d799","Type":"ContainerStarted","Data":"de84ea928c21e7d851428ee60843720d656101db59c03c12935d28595c4b6525"} Mar 12 13:28:05 crc kubenswrapper[4778]: I0312 13:28:05.104723 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:28:05 crc kubenswrapper[4778]: I0312 13:28:05.121658 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bc06ca-f4e6-4fde-bd5d-882714d9652c-cert\") pod \"infra-operator-controller-manager-5995f4446f-5d6qz\" (UID: \"02bc06ca-f4e6-4fde-bd5d-882714d9652c\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:28:05 crc kubenswrapper[4778]: I0312 13:28:05.156749 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:28:07 crc kubenswrapper[4778]: I0312 13:28:07.379310 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:28:07 crc kubenswrapper[4778]: I0312 13:28:07.379866 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:07 crc kubenswrapper[4778]: I0312 13:28:07.379915 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:07 crc kubenswrapper[4778]: E0312 13:28:07.380815 4778 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 13:28:07 crc kubenswrapper[4778]: E0312 13:28:07.382344 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs podName:d0784623-5f08-4109-9c7e-0a329210ce07 nodeName:}" failed. No retries permitted until 2026-03-12 13:28:23.382294431 +0000 UTC m=+1121.830989847 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs") pod "openstack-operator-controller-manager-5785b7957-7vdgw" (UID: "d0784623-5f08-4109-9c7e-0a329210ce07") : secret "webhook-server-cert" not found Mar 12 13:28:07 crc kubenswrapper[4778]: I0312 13:28:07.393761 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-metrics-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:07 crc kubenswrapper[4778]: I0312 13:28:07.440814 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7d316e-6896-4f84-8423-6f79778c1c6b-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6\" (UID: \"4f7d316e-6896-4f84-8423-6f79778c1c6b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:28:07 crc kubenswrapper[4778]: I0312 13:28:07.684712 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:28:12 crc kubenswrapper[4778]: E0312 13:28:12.716062 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:571f369855b0891a2b14e54a4c1c5ae2fbbd5de4c8fddd48e81033aad4b26423" Mar 12 13:28:12 crc kubenswrapper[4778]: E0312 13:28:12.716714 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:571f369855b0891a2b14e54a4c1c5ae2fbbd5de4c8fddd48e81033aad4b26423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dvw2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-677bd678f7-6h2c2_openstack-operators(ffb8a1f4-4533-4368-a900-95d37fe1d3ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:12 crc kubenswrapper[4778]: E0312 13:28:12.718227 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" podUID="ffb8a1f4-4533-4368-a900-95d37fe1d3ad" Mar 12 13:28:13 crc kubenswrapper[4778]: E0312 13:28:13.068211 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:571f369855b0891a2b14e54a4c1c5ae2fbbd5de4c8fddd48e81033aad4b26423\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" podUID="ffb8a1f4-4533-4368-a900-95d37fe1d3ad" Mar 12 13:28:17 crc kubenswrapper[4778]: E0312 13:28:17.218867 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6" Mar 12 13:28:17 crc kubenswrapper[4778]: E0312 13:28:17.219445 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4r2pb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6d9d6b584d-4jgt8_openstack-operators(4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:17 crc kubenswrapper[4778]: E0312 13:28:17.220627 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" podUID="4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.023385 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.023608 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g8jj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-77b6666d85-b7tkm_openstack-operators(e290c1ea-a39d-451e-a24b-17a2b61ff6f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.025032 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" podUID="e290c1ea-a39d-451e-a24b-17a2b61ff6f0" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.106416 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6\\\"\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" podUID="e290c1ea-a39d-451e-a24b-17a2b61ff6f0" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.106909 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" podUID="4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.883658 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.883899 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j7lr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-66d56f6ff4-9n6jv_openstack-operators(ad531191-d7c5-4ef6-9929-3a5869751d98): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:18 crc kubenswrapper[4778]: E0312 13:28:18.885013 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" podUID="ad531191-d7c5-4ef6-9929-3a5869751d98" Mar 12 13:28:19 crc kubenswrapper[4778]: E0312 13:28:19.109156 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7\\\"\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" podUID="ad531191-d7c5-4ef6-9929-3a5869751d98" Mar 12 13:28:19 crc kubenswrapper[4778]: E0312 13:28:19.625901 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571" Mar 12 13:28:19 crc kubenswrapper[4778]: E0312 13:28:19.626415 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5sl2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5f4f55cb5c-cdgg9_openstack-operators(1a01d06c-be6f-45de-a22d-c8f1058a3a84): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:19 crc kubenswrapper[4778]: E0312 13:28:19.627690 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" podUID="1a01d06c-be6f-45de-a22d-c8f1058a3a84" Mar 12 13:28:20 crc kubenswrapper[4778]: E0312 13:28:20.155370 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" podUID="1a01d06c-be6f-45de-a22d-c8f1058a3a84" Mar 12 13:28:20 crc kubenswrapper[4778]: E0312 13:28:20.333150 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:7c0da25380c91ffd1940d75eaa71b6842a6a4cf4056e62d6b0d237897b74e4d9" Mar 12 13:28:20 crc kubenswrapper[4778]: E0312 13:28:20.333435 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:7c0da25380c91ffd1940d75eaa71b6842a6a4cf4056e62d6b0d237897b74e4d9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ljqr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-984cd4dcf-xm4cc_openstack-operators(c8818ac0-af8b-42c9-a923-425fe79ed203): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:20 crc kubenswrapper[4778]: E0312 13:28:20.334642 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" podUID="c8818ac0-af8b-42c9-a923-425fe79ed203" Mar 12 13:28:21 crc kubenswrapper[4778]: E0312 13:28:21.009692 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554" Mar 12 13:28:21 crc kubenswrapper[4778]: E0312 13:28:21.009942 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j7fsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6dd88c6f67-2tjsk_openstack-operators(8c02ecb8-0e15-4672-823a-c4437ca5bf8c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:21 crc kubenswrapper[4778]: E0312 13:28:21.011168 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" podUID="8c02ecb8-0e15-4672-823a-c4437ca5bf8c" Mar 12 13:28:21 crc kubenswrapper[4778]: E0312 13:28:21.162640 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" podUID="8c02ecb8-0e15-4672-823a-c4437ca5bf8c" Mar 12 13:28:21 crc kubenswrapper[4778]: E0312 13:28:21.163956 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:7c0da25380c91ffd1940d75eaa71b6842a6a4cf4056e62d6b0d237897b74e4d9\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" podUID="c8818ac0-af8b-42c9-a923-425fe79ed203" Mar 12 13:28:22 crc kubenswrapper[4778]: E0312 13:28:22.140868 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978" Mar 12 13:28:22 crc kubenswrapper[4778]: E0312 13:28:22.141129 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-764zg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-wvpf8_openstack-operators(52524252-25bd-49e5-822e-3d4668aff2f9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:22 crc kubenswrapper[4778]: E0312 13:28:22.142409 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" podUID="52524252-25bd-49e5-822e-3d4668aff2f9" Mar 12 13:28:22 crc kubenswrapper[4778]: E0312 13:28:22.167923 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" podUID="52524252-25bd-49e5-822e-3d4668aff2f9" Mar 12 13:28:23 crc kubenswrapper[4778]: E0312 13:28:23.339924 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4" Mar 12 13:28:23 crc kubenswrapper[4778]: E0312 13:28:23.340664 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wb75x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-68f45f9d9f-pn8tk_openstack-operators(5e38a4fd-95f8-437b-923b-eca33b1387e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:23 crc kubenswrapper[4778]: E0312 13:28:23.341970 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" podUID="5e38a4fd-95f8-437b-923b-eca33b1387e6" Mar 12 13:28:23 crc kubenswrapper[4778]: I0312 13:28:23.400393 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:23 crc kubenswrapper[4778]: I0312 13:28:23.407658 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d0784623-5f08-4109-9c7e-0a329210ce07-webhook-certs\") pod \"openstack-operator-controller-manager-5785b7957-7vdgw\" (UID: \"d0784623-5f08-4109-9c7e-0a329210ce07\") " pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:23 crc kubenswrapper[4778]: I0312 13:28:23.707768 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:24 crc kubenswrapper[4778]: E0312 13:28:24.244924 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4\\\"\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" podUID="5e38a4fd-95f8-437b-923b-eca33b1387e6" Mar 12 13:28:24 crc kubenswrapper[4778]: E0312 13:28:24.592168 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f" Mar 12 13:28:24 crc kubenswrapper[4778]: E0312 13:28:24.592521 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mj8wf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6bbb499bbc-qb8s8_openstack-operators(98a4cfbd-3037-48b5-9047-5d574dcc0aca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:24 crc kubenswrapper[4778]: E0312 13:28:24.593867 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" podUID="98a4cfbd-3037-48b5-9047-5d574dcc0aca" Mar 12 13:28:25 crc kubenswrapper[4778]: E0312 13:28:25.264567 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" podUID="98a4cfbd-3037-48b5-9047-5d574dcc0aca" Mar 12 13:28:25 crc kubenswrapper[4778]: E0312 13:28:25.563108 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f" Mar 12 13:28:25 crc kubenswrapper[4778]: E0312 13:28:25.563446 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ktk6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bbc5b68f9-bbgmb_openstack-operators(8d38fd7e-6fa1-4b0c-9c82-9c57290c7837): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:25 crc kubenswrapper[4778]: E0312 13:28:25.564789 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" podUID="8d38fd7e-6fa1-4b0c-9c82-9c57290c7837" Mar 12 13:28:26 crc kubenswrapper[4778]: I0312 13:28:26.255873 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:28:26 crc kubenswrapper[4778]: E0312 13:28:26.267980 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" podUID="8d38fd7e-6fa1-4b0c-9c82-9c57290c7837" Mar 12 13:28:28 crc kubenswrapper[4778]: E0312 13:28:28.048144 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721" Mar 12 13:28:28 crc kubenswrapper[4778]: E0312 13:28:28.048671 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-txndm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-dd2ft_openstack-operators(076835c9-352b-4e40-80c4-3bce3bb80594): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:28 crc kubenswrapper[4778]: E0312 13:28:28.049898 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" podUID="076835c9-352b-4e40-80c4-3bce3bb80594" Mar 12 13:28:28 crc kubenswrapper[4778]: E0312 13:28:28.303643 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" podUID="076835c9-352b-4e40-80c4-3bce3bb80594" Mar 12 13:28:28 crc kubenswrapper[4778]: I0312 13:28:28.558240 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:28:28 crc kubenswrapper[4778]: I0312 13:28:28.558305 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:28:29 crc kubenswrapper[4778]: E0312 13:28:29.079850 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:27c84b712abc2df6108e22636075eec25fea0229800f38594a492fd41b02c49d" Mar 12 13:28:29 crc kubenswrapper[4778]: E0312 13:28:29.080044 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:27c84b712abc2df6108e22636075eec25fea0229800f38594a492fd41b02c49d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d2ql6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6cd66dbd4b-gfv5z_openstack-operators(6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:29 crc kubenswrapper[4778]: E0312 13:28:29.081295 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" podUID="6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c" Mar 12 13:28:29 crc kubenswrapper[4778]: E0312 13:28:29.309078 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:27c84b712abc2df6108e22636075eec25fea0229800f38594a492fd41b02c49d\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" podUID="6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.105576 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.106860 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ldr22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-shf7b_openstack-operators(034f39d8-a33e-4e37-bcde-51fb22debdd1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.108106 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" podUID="034f39d8-a33e-4e37-bcde-51fb22debdd1" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.319082 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" podUID="034f39d8-a33e-4e37-bcde-51fb22debdd1" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.595050 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.595256 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bjmh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-684f77d66d-7dxdh_openstack-operators(7e02c37f-b9af-46c9-a743-03ead9b060db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.596456 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" podUID="7e02c37f-b9af-46c9-a743-03ead9b060db" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.684825 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.97:5001/openstack-k8s-operators/nova-operator:8734adf928be66fa1f808466edcc3ea058f7094f" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.685538 4778 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.97:5001/openstack-k8s-operators/nova-operator:8734adf928be66fa1f808466edcc3ea058f7094f" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.685711 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.129.56.97:5001/openstack-k8s-operators/nova-operator:8734adf928be66fa1f808466edcc3ea058f7094f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qw6mz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-686d5f9fbd-vv9rc_openstack-operators(d7288cc6-4247-4d03-bd37-9862243bf613): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:28:30 crc kubenswrapper[4778]: E0312 13:28:30.687332 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" podUID="d7288cc6-4247-4d03-bd37-9862243bf613" Mar 12 13:28:30 crc kubenswrapper[4778]: I0312 13:28:30.984541 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw"] Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.124917 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz"] Mar 12 13:28:31 crc kubenswrapper[4778]: W0312 13:28:31.138907 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02bc06ca_f4e6_4fde_bd5d_882714d9652c.slice/crio-f4773d1227aea67668cd4cfcd9007b1984af5d83edd8146ec84d606804c93441 WatchSource:0}: Error finding container f4773d1227aea67668cd4cfcd9007b1984af5d83edd8146ec84d606804c93441: Status 404 returned error can't find the container with id f4773d1227aea67668cd4cfcd9007b1984af5d83edd8146ec84d606804c93441 Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.248638 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6"] Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.323953 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" event={"ID":"20d587ee-7b57-4b99-a800-c6d46322d799","Type":"ContainerStarted","Data":"c06e4e1b6c58e04407e154a6eb32ce96d2dfbf0e7e2f81409f2e784cc2f29542"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.325685 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" event={"ID":"db7f6b97-2903-44bf-803f-c00c337400b9","Type":"ContainerStarted","Data":"a5bf5ec21d4065da295f00d28a5287cd362a01c304f74d7e5512faa219b3d7de"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.325866 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.330324 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" event={"ID":"ed9b9271-4ae9-440a-9411-15d46267106e","Type":"ContainerStarted","Data":"7b17321aaf7cd993b288c637ad0612dc4c574dab465fc6e4b6f72db4a3b80c18"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.330389 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.331512 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" event={"ID":"4f7d316e-6896-4f84-8423-6f79778c1c6b","Type":"ContainerStarted","Data":"adb2d40e89caf9534d5c4dabf505feb3801f9b44831b0c55adf126795bfad3bd"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.333097 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" event={"ID":"02bc06ca-f4e6-4fde-bd5d-882714d9652c","Type":"ContainerStarted","Data":"f4773d1227aea67668cd4cfcd9007b1984af5d83edd8146ec84d606804c93441"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.334418 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" event={"ID":"d0784623-5f08-4109-9c7e-0a329210ce07","Type":"ContainerStarted","Data":"d33410a9882018e414468638661548551ebef7d157616bca16c7e7be28aced14"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.334441 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" event={"ID":"d0784623-5f08-4109-9c7e-0a329210ce07","Type":"ContainerStarted","Data":"5f01fcfb5d6c209379adeb47c56eb146cd7d15d24ce620cacec56510e790e55e"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.335067 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.336503 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" event={"ID":"ffb8a1f4-4533-4368-a900-95d37fe1d3ad","Type":"ContainerStarted","Data":"7bb3ac22f2df17855f0d4f1ce9050828bca873c51e045b0423f43e2e9b38806c"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.336999 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.338259 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" event={"ID":"2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f","Type":"ContainerStarted","Data":"0502ba772b336b3388415a546413a2b0363d7512fe9d7157ee6e2c17da947864"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.338589 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.340842 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" event={"ID":"64a36384-f2e6-4077-b2ca-de2a6ce6ea06","Type":"ContainerStarted","Data":"d44f9cc5c66a90f293068c46be5aaf2887ded95f7ed413336a8e90b5043e9a57"} Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.340866 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:28:31 crc kubenswrapper[4778]: E0312 13:28:31.341993 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" podUID="7e02c37f-b9af-46c9-a743-03ead9b060db" Mar 12 13:28:31 crc kubenswrapper[4778]: E0312 13:28:31.342373 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.97:5001/openstack-k8s-operators/nova-operator:8734adf928be66fa1f808466edcc3ea058f7094f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" podUID="d7288cc6-4247-4d03-bd37-9862243bf613" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.351092 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" podStartSLOduration=2.380480556 podStartE2EDuration="31.351070239s" podCreationTimestamp="2026-03-12 13:28:00 +0000 UTC" firstStartedPulling="2026-03-12 13:28:01.70172612 +0000 UTC m=+1100.150421506" lastFinishedPulling="2026-03-12 13:28:30.672315793 +0000 UTC m=+1129.121011189" observedRunningTime="2026-03-12 13:28:31.344244494 +0000 UTC m=+1129.792939910" watchObservedRunningTime="2026-03-12 13:28:31.351070239 +0000 UTC m=+1129.799765635" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.363459 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" podStartSLOduration=4.6482561 podStartE2EDuration="44.363439491s" podCreationTimestamp="2026-03-12 13:27:47 +0000 UTC" firstStartedPulling="2026-03-12 13:27:50.957973946 +0000 UTC m=+1089.406669342" lastFinishedPulling="2026-03-12 13:28:30.673157337 +0000 UTC m=+1129.121852733" observedRunningTime="2026-03-12 13:28:31.362469623 +0000 UTC m=+1129.811165029" watchObservedRunningTime="2026-03-12 13:28:31.363439491 +0000 UTC m=+1129.812134877" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.410713 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" podStartSLOduration=7.685833703 podStartE2EDuration="42.410698728s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.894567489 +0000 UTC m=+1093.343262885" lastFinishedPulling="2026-03-12 13:28:29.619432514 +0000 UTC m=+1128.068127910" observedRunningTime="2026-03-12 13:28:31.409099333 +0000 UTC m=+1129.857794729" watchObservedRunningTime="2026-03-12 13:28:31.410698728 +0000 UTC m=+1129.859394124" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.843075 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" podStartSLOduration=41.843051531 podStartE2EDuration="41.843051531s" podCreationTimestamp="2026-03-12 13:27:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:28:31.495383402 +0000 UTC m=+1129.944078798" watchObservedRunningTime="2026-03-12 13:28:31.843051531 +0000 UTC m=+1130.291746927" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.843609 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" podStartSLOduration=6.6207949379999995 podStartE2EDuration="43.843600507s" podCreationTimestamp="2026-03-12 13:27:48 +0000 UTC" firstStartedPulling="2026-03-12 13:27:52.858774027 +0000 UTC m=+1091.307469423" lastFinishedPulling="2026-03-12 13:28:30.081579596 +0000 UTC m=+1128.530274992" observedRunningTime="2026-03-12 13:28:31.840835018 +0000 UTC m=+1130.289530424" watchObservedRunningTime="2026-03-12 13:28:31.843600507 +0000 UTC m=+1130.292295913" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.860267 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" podStartSLOduration=7.530400835 podStartE2EDuration="42.860244441s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:55.243491156 +0000 UTC m=+1093.692186552" lastFinishedPulling="2026-03-12 13:28:30.573334762 +0000 UTC m=+1129.022030158" observedRunningTime="2026-03-12 13:28:31.858392948 +0000 UTC m=+1130.307088334" watchObservedRunningTime="2026-03-12 13:28:31.860244441 +0000 UTC m=+1130.308939837" Mar 12 13:28:31 crc kubenswrapper[4778]: I0312 13:28:31.876867 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" podStartSLOduration=8.115836251 podStartE2EDuration="42.876846804s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:55.319701698 +0000 UTC m=+1093.768397094" lastFinishedPulling="2026-03-12 13:28:30.080712251 +0000 UTC m=+1128.529407647" observedRunningTime="2026-03-12 13:28:31.875323171 +0000 UTC m=+1130.324018587" watchObservedRunningTime="2026-03-12 13:28:31.876846804 +0000 UTC m=+1130.325542200" Mar 12 13:28:32 crc kubenswrapper[4778]: I0312 13:28:32.349278 4778 generic.go:334] "Generic (PLEG): container finished" podID="20d587ee-7b57-4b99-a800-c6d46322d799" containerID="c06e4e1b6c58e04407e154a6eb32ce96d2dfbf0e7e2f81409f2e784cc2f29542" exitCode=0 Mar 12 13:28:32 crc kubenswrapper[4778]: I0312 13:28:32.351374 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" event={"ID":"20d587ee-7b57-4b99-a800-c6d46322d799","Type":"ContainerDied","Data":"c06e4e1b6c58e04407e154a6eb32ce96d2dfbf0e7e2f81409f2e784cc2f29542"} Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.302596 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.365141 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" event={"ID":"20d587ee-7b57-4b99-a800-c6d46322d799","Type":"ContainerDied","Data":"de84ea928c21e7d851428ee60843720d656101db59c03c12935d28595c4b6525"} Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.365204 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de84ea928c21e7d851428ee60843720d656101db59c03c12935d28595c4b6525" Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.365271 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555368-d2cpg" Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.415723 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555362-hlj7f"] Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.421133 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555362-hlj7f"] Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.421147 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m2kn\" (UniqueName: \"kubernetes.io/projected/20d587ee-7b57-4b99-a800-c6d46322d799-kube-api-access-9m2kn\") pod \"20d587ee-7b57-4b99-a800-c6d46322d799\" (UID: \"20d587ee-7b57-4b99-a800-c6d46322d799\") " Mar 12 13:28:34 crc kubenswrapper[4778]: I0312 13:28:34.939048 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d587ee-7b57-4b99-a800-c6d46322d799-kube-api-access-9m2kn" (OuterVolumeSpecName: "kube-api-access-9m2kn") pod "20d587ee-7b57-4b99-a800-c6d46322d799" (UID: "20d587ee-7b57-4b99-a800-c6d46322d799"). InnerVolumeSpecName "kube-api-access-9m2kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:28:35 crc kubenswrapper[4778]: I0312 13:28:35.039659 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m2kn\" (UniqueName: \"kubernetes.io/projected/20d587ee-7b57-4b99-a800-c6d46322d799-kube-api-access-9m2kn\") on node \"crc\" DevicePath \"\"" Mar 12 13:28:36 crc kubenswrapper[4778]: I0312 13:28:36.278068 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b" path="/var/lib/kubelet/pods/9da11ea2-3173-4f25-8f0e-3ccc5a0ca18b/volumes" Mar 12 13:28:38 crc kubenswrapper[4778]: I0312 13:28:38.473039 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6h2c2" Mar 12 13:28:39 crc kubenswrapper[4778]: I0312 13:28:39.252731 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gknp2" Mar 12 13:28:40 crc kubenswrapper[4778]: I0312 13:28:40.282684 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-jlbft" Mar 12 13:28:41 crc kubenswrapper[4778]: I0312 13:28:41.075537 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-677c674df7-84mps" Mar 12 13:28:41 crc kubenswrapper[4778]: I0312 13:28:41.538839 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-pcfrz" Mar 12 13:28:43 crc kubenswrapper[4778]: I0312 13:28:43.763864 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5785b7957-7vdgw" Mar 12 13:28:47 crc kubenswrapper[4778]: I0312 13:28:47.493706 4778 scope.go:117] "RemoveContainer" containerID="1c6932f83080c12204b2bc10f63ca97fbee0fb238358dc69be9a27d4fc46a8a5" Mar 12 13:28:58 crc kubenswrapper[4778]: I0312 13:28:58.557718 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:28:58 crc kubenswrapper[4778]: I0312 13:28:58.558383 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:29:41 crc kubenswrapper[4778]: I0312 13:29:28.558090 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:29:41 crc kubenswrapper[4778]: I0312 13:29:28.559154 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:29:41 crc kubenswrapper[4778]: I0312 13:29:28.559255 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:29:41 crc kubenswrapper[4778]: I0312 13:29:28.560273 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b4b372cac8f288fc2585670d5ab7c00c41331f173130d39b164aa74e4e3e398"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:29:41 crc kubenswrapper[4778]: I0312 13:29:28.560337 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://3b4b372cac8f288fc2585670d5ab7c00c41331f173130d39b164aa74e4e3e398" gracePeriod=600 Mar 12 13:29:41 crc kubenswrapper[4778]: E0312 13:29:41.546272 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:417a4ede6dce5d088ce7dc1ac6e9dab30f3b532bd5b506e2df65d6eaecbc7cb9" Mar 12 13:29:41 crc kubenswrapper[4778]: E0312 13:29:41.547491 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:417a4ede6dce5d088ce7dc1ac6e9dab30f3b532bd5b506e2df65d6eaecbc7cb9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpx49,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-5995f4446f-5d6qz_openstack-operators(02bc06ca-f4e6-4fde-bd5d-882714d9652c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:29:41 crc kubenswrapper[4778]: E0312 13:29:41.548661 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" podUID="02bc06ca-f4e6-4fde-bd5d-882714d9652c" Mar 12 13:29:52 crc kubenswrapper[4778]: I0312 13:29:42.446103 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"3b4b372cac8f288fc2585670d5ab7c00c41331f173130d39b164aa74e4e3e398"} Mar 12 13:29:52 crc kubenswrapper[4778]: I0312 13:29:42.446136 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="3b4b372cac8f288fc2585670d5ab7c00c41331f173130d39b164aa74e4e3e398" exitCode=0 Mar 12 13:29:52 crc kubenswrapper[4778]: I0312 13:29:42.446212 4778 scope.go:117] "RemoveContainer" containerID="b65e287d42eea6146877a35b0789c26ac0ef9f5d251a760b59f08b3fef055d65" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.154448 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f"] Mar 12 13:30:00 crc kubenswrapper[4778]: E0312 13:30:00.155394 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d587ee-7b57-4b99-a800-c6d46322d799" containerName="oc" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.155412 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d587ee-7b57-4b99-a800-c6d46322d799" containerName="oc" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.155586 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d587ee-7b57-4b99-a800-c6d46322d799" containerName="oc" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.156209 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.159769 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.161594 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.164790 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555370-6zrgd"] Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.165926 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.168539 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.168718 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.168730 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.171928 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555370-6zrgd"] Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.178716 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f"] Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.281469 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ks2n\" (UniqueName: \"kubernetes.io/projected/1c682acb-240b-44d4-a2be-0ea0cd913af1-kube-api-access-9ks2n\") pod \"auto-csr-approver-29555370-6zrgd\" (UID: \"1c682acb-240b-44d4-a2be-0ea0cd913af1\") " pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.281725 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xsfh\" (UniqueName: \"kubernetes.io/projected/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-kube-api-access-4xsfh\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.281830 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-secret-volume\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.281992 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-config-volume\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.383545 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-config-volume\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.383930 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ks2n\" (UniqueName: \"kubernetes.io/projected/1c682acb-240b-44d4-a2be-0ea0cd913af1-kube-api-access-9ks2n\") pod \"auto-csr-approver-29555370-6zrgd\" (UID: \"1c682acb-240b-44d4-a2be-0ea0cd913af1\") " pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.384049 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xsfh\" (UniqueName: \"kubernetes.io/projected/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-kube-api-access-4xsfh\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.384201 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-secret-volume\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.384882 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-config-volume\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.391637 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-secret-volume\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.404348 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ks2n\" (UniqueName: \"kubernetes.io/projected/1c682acb-240b-44d4-a2be-0ea0cd913af1-kube-api-access-9ks2n\") pod \"auto-csr-approver-29555370-6zrgd\" (UID: \"1c682acb-240b-44d4-a2be-0ea0cd913af1\") " pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.413718 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xsfh\" (UniqueName: \"kubernetes.io/projected/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-kube-api-access-4xsfh\") pod \"collect-profiles-29555370-zcp5f\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.491919 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:00 crc kubenswrapper[4778]: I0312 13:30:00.504242 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:10 crc kubenswrapper[4778]: I0312 13:30:10.704251 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f"] Mar 12 13:30:10 crc kubenswrapper[4778]: I0312 13:30:10.789435 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555370-6zrgd"] Mar 12 13:30:10 crc kubenswrapper[4778]: W0312 13:30:10.804559 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bf03685_d980_41f0_bbc5_84b9ae0ce1df.slice/crio-1a5e91c9d02c7de2c4009f12029cabe072d8d04d8718617498c0566b1d22e0df WatchSource:0}: Error finding container 1a5e91c9d02c7de2c4009f12029cabe072d8d04d8718617498c0566b1d22e0df: Status 404 returned error can't find the container with id 1a5e91c9d02c7de2c4009f12029cabe072d8d04d8718617498c0566b1d22e0df Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.688486 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" event={"ID":"8bf03685-d980-41f0-bbc5-84b9ae0ce1df","Type":"ContainerStarted","Data":"fa067a709ad1af5d5b9327929891ffc04839dd2d8aba3cc70c48dbfeabd353b9"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.689056 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" event={"ID":"8bf03685-d980-41f0-bbc5-84b9ae0ce1df","Type":"ContainerStarted","Data":"1a5e91c9d02c7de2c4009f12029cabe072d8d04d8718617498c0566b1d22e0df"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.708428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" event={"ID":"1a01d06c-be6f-45de-a22d-c8f1058a3a84","Type":"ContainerStarted","Data":"682b989352e7cb03a00ffe09b099d2da7b43ddf31a4c0e296a43178b62c7b528"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.709056 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.717994 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" podStartSLOduration=11.717976851 podStartE2EDuration="11.717976851s" podCreationTimestamp="2026-03-12 13:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:30:11.716300803 +0000 UTC m=+1230.164996199" watchObservedRunningTime="2026-03-12 13:30:11.717976851 +0000 UTC m=+1230.166672247" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.738431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" event={"ID":"5e38a4fd-95f8-437b-923b-eca33b1387e6","Type":"ContainerStarted","Data":"a82c48c3a9e4df5db23c33aa9590f8e780acf018ad62dbbe99c04313660852fc"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.739077 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.773803 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" event={"ID":"8d38fd7e-6fa1-4b0c-9c82-9c57290c7837","Type":"ContainerStarted","Data":"d13442b4e35e6c42e7a2d19c65808770c225b101c756ff5c9b92cd147f000999"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.774792 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.790433 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"572aad6c3b1a3f7c9ef45b8b4feb0d367e7e7916d0ab8dd064e2b8ac87268c51"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.797915 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" podStartSLOduration=7.417947677 podStartE2EDuration="2m22.797891828s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.552225511 +0000 UTC m=+1093.000920897" lastFinishedPulling="2026-03-12 13:30:09.932169612 +0000 UTC m=+1228.380865048" observedRunningTime="2026-03-12 13:30:11.74356703 +0000 UTC m=+1230.192262426" watchObservedRunningTime="2026-03-12 13:30:11.797891828 +0000 UTC m=+1230.246587224" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.799610 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" podStartSLOduration=7.095459163 podStartE2EDuration="2m22.799602367s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.105256569 +0000 UTC m=+1092.553951965" lastFinishedPulling="2026-03-12 13:30:09.809399763 +0000 UTC m=+1228.258095169" observedRunningTime="2026-03-12 13:30:11.785351831 +0000 UTC m=+1230.234047237" watchObservedRunningTime="2026-03-12 13:30:11.799602367 +0000 UTC m=+1230.248297773" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.809233 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" event={"ID":"4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71","Type":"ContainerStarted","Data":"944e9b4f5593188c7d0547feee17bc4fdf087537a72c1068daf9af76b28418a3"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.809967 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.819036 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" event={"ID":"52524252-25bd-49e5-822e-3d4668aff2f9","Type":"ContainerStarted","Data":"1dedc528009aad1b174fe4409306d7b650583ab562c684d3e834b3887a83ce9c"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.821142 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" event={"ID":"076835c9-352b-4e40-80c4-3bce3bb80594","Type":"ContainerStarted","Data":"9573cdd27cbfe1c6f23115e81f3555e406dcbbe6c3b3fbf03eb74d488e7d2632"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.821214 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" podStartSLOduration=7.42015443 podStartE2EDuration="2m22.820858153s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.851175811 +0000 UTC m=+1093.299871217" lastFinishedPulling="2026-03-12 13:30:10.251879554 +0000 UTC m=+1228.700574940" observedRunningTime="2026-03-12 13:30:11.816336934 +0000 UTC m=+1230.265032330" watchObservedRunningTime="2026-03-12 13:30:11.820858153 +0000 UTC m=+1230.269553549" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.821587 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.822197 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" event={"ID":"98a4cfbd-3037-48b5-9047-5d574dcc0aca","Type":"ContainerStarted","Data":"fe9764764913797541ce93765c0193b347c66bf42c36d8f4e18119306c3ba418"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.825595 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.834990 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" event={"ID":"02bc06ca-f4e6-4fde-bd5d-882714d9652c","Type":"ContainerStarted","Data":"18af05ec7a686ea1e7d1086f5ea04254d6c6ca199fe137a46915b1da7f1fc180"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.835204 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.835901 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" event={"ID":"1c682acb-240b-44d4-a2be-0ea0cd913af1","Type":"ContainerStarted","Data":"623f846d223ee24a6cba599fe831dbd5bf60da2ddff32e7acc90057e22b71876"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.839820 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" podStartSLOduration=7.1148405 podStartE2EDuration="2m23.839806703s" podCreationTimestamp="2026-03-12 13:27:48 +0000 UTC" firstStartedPulling="2026-03-12 13:27:52.271251694 +0000 UTC m=+1090.719947090" lastFinishedPulling="2026-03-12 13:30:08.996217897 +0000 UTC m=+1227.444913293" observedRunningTime="2026-03-12 13:30:11.834753529 +0000 UTC m=+1230.283448925" watchObservedRunningTime="2026-03-12 13:30:11.839806703 +0000 UTC m=+1230.288502089" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.845448 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" event={"ID":"034f39d8-a33e-4e37-bcde-51fb22debdd1","Type":"ContainerStarted","Data":"07692534a3d4f2d2a10f9b6c0d2df8a72199cae2af3442e8e487917a6698768b"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.867890 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" event={"ID":"c8818ac0-af8b-42c9-a923-425fe79ed203","Type":"ContainerStarted","Data":"94e31167ba37ebbdf9a654d9e583c8e77901521d8cacc623858135237eaafcb5"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.868851 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.897849 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" event={"ID":"8c02ecb8-0e15-4672-823a-c4437ca5bf8c","Type":"ContainerStarted","Data":"1bcf3d1a36b393c0d7a06b92fe96b249521ebb9c61ddf71fe57200c7ee9a1a3f"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.898805 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.901993 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" podStartSLOduration=44.6801485 podStartE2EDuration="2m23.901974915s" podCreationTimestamp="2026-03-12 13:27:48 +0000 UTC" firstStartedPulling="2026-03-12 13:28:31.14586455 +0000 UTC m=+1129.594559966" lastFinishedPulling="2026-03-12 13:30:10.367690965 +0000 UTC m=+1228.816386381" observedRunningTime="2026-03-12 13:30:11.899280068 +0000 UTC m=+1230.347975464" watchObservedRunningTime="2026-03-12 13:30:11.901974915 +0000 UTC m=+1230.350670311" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.943446 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" event={"ID":"6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c","Type":"ContainerStarted","Data":"e9665833f856db66cbdf9ceafdbd4ac9eb165892b060faf205d90b6dd7e9f1c7"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.943732 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.967403 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" event={"ID":"d7288cc6-4247-4d03-bd37-9862243bf613","Type":"ContainerStarted","Data":"3fc68162292e71c8fc5f9cbe96d0444169057f2d6a017a0fc30ef68581d2f893"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.968023 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.974040 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" podStartSLOduration=6.6399989999999995 podStartE2EDuration="2m22.974021138s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.037977 +0000 UTC m=+1092.486672396" lastFinishedPulling="2026-03-12 13:30:10.371999128 +0000 UTC m=+1228.820694534" observedRunningTime="2026-03-12 13:30:11.943981502 +0000 UTC m=+1230.392676898" watchObservedRunningTime="2026-03-12 13:30:11.974021138 +0000 UTC m=+1230.422716534" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.980727 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" podStartSLOduration=7.901207821 podStartE2EDuration="2m22.980698559s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.851607513 +0000 UTC m=+1093.300302909" lastFinishedPulling="2026-03-12 13:30:09.931098251 +0000 UTC m=+1228.379793647" observedRunningTime="2026-03-12 13:30:11.972471884 +0000 UTC m=+1230.421167280" watchObservedRunningTime="2026-03-12 13:30:11.980698559 +0000 UTC m=+1230.429393955" Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.991872 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" event={"ID":"ad531191-d7c5-4ef6-9929-3a5869751d98","Type":"ContainerStarted","Data":"608055fcdf6f010cf5a045ff56b1d3ff4a839075052dae92af32dd512d1a10a8"} Mar 12 13:30:11 crc kubenswrapper[4778]: I0312 13:30:11.993121 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.003509 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" event={"ID":"7e02c37f-b9af-46c9-a743-03ead9b060db","Type":"ContainerStarted","Data":"3070f4c0343eea3d825bef783dce2823c5ca2fb514d1402fcf5f63c0219e7ac7"} Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.004345 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.007608 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-shf7b" podStartSLOduration=5.980794323 podStartE2EDuration="2m21.007573745s" podCreationTimestamp="2026-03-12 13:27:51 +0000 UTC" firstStartedPulling="2026-03-12 13:27:55.342519039 +0000 UTC m=+1093.791214435" lastFinishedPulling="2026-03-12 13:30:10.369298461 +0000 UTC m=+1228.817993857" observedRunningTime="2026-03-12 13:30:12.000832523 +0000 UTC m=+1230.449527929" watchObservedRunningTime="2026-03-12 13:30:12.007573745 +0000 UTC m=+1230.456269141" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.023415 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" event={"ID":"4f7d316e-6896-4f84-8423-6f79778c1c6b","Type":"ContainerStarted","Data":"49e66e3e82cd4589f6d0c32156c3c77fcc2471f8841861dbcbab5a2ef92bd057"} Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.024168 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.064808 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" event={"ID":"e290c1ea-a39d-451e-a24b-17a2b61ff6f0","Type":"ContainerStarted","Data":"ba762064366842b2ae297d67ef6885f977d2d867bb935f92964f677eda3bc8b1"} Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.065845 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.094465 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" podStartSLOduration=6.763406299 podStartE2EDuration="2m23.094423731s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.038578297 +0000 UTC m=+1092.487273703" lastFinishedPulling="2026-03-12 13:30:10.369595729 +0000 UTC m=+1228.818291135" observedRunningTime="2026-03-12 13:30:12.087547455 +0000 UTC m=+1230.536242851" watchObservedRunningTime="2026-03-12 13:30:12.094423731 +0000 UTC m=+1230.543119117" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.144856 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" podStartSLOduration=7.684919612 podStartE2EDuration="2m24.144831707s" podCreationTimestamp="2026-03-12 13:27:48 +0000 UTC" firstStartedPulling="2026-03-12 13:27:53.349509898 +0000 UTC m=+1091.798205294" lastFinishedPulling="2026-03-12 13:30:09.809421983 +0000 UTC m=+1228.258117389" observedRunningTime="2026-03-12 13:30:12.049661004 +0000 UTC m=+1230.498356400" watchObservedRunningTime="2026-03-12 13:30:12.144831707 +0000 UTC m=+1230.593527103" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.148967 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" podStartSLOduration=7.618969517 podStartE2EDuration="2m23.148952474s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:27:54.897952645 +0000 UTC m=+1093.346648031" lastFinishedPulling="2026-03-12 13:30:10.427935602 +0000 UTC m=+1228.876630988" observedRunningTime="2026-03-12 13:30:12.118695202 +0000 UTC m=+1230.567390588" watchObservedRunningTime="2026-03-12 13:30:12.148952474 +0000 UTC m=+1230.597647870" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.169485 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" podStartSLOduration=7.505351563 podStartE2EDuration="2m22.169453929s" podCreationTimestamp="2026-03-12 13:27:50 +0000 UTC" firstStartedPulling="2026-03-12 13:27:55.267968503 +0000 UTC m=+1093.716663899" lastFinishedPulling="2026-03-12 13:30:09.932070829 +0000 UTC m=+1228.380766265" observedRunningTime="2026-03-12 13:30:12.14527384 +0000 UTC m=+1230.593969256" watchObservedRunningTime="2026-03-12 13:30:12.169453929 +0000 UTC m=+1230.618149395" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.207364 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" podStartSLOduration=44.898559297 podStartE2EDuration="2m23.207344289s" podCreationTimestamp="2026-03-12 13:27:49 +0000 UTC" firstStartedPulling="2026-03-12 13:28:31.264531902 +0000 UTC m=+1129.713227298" lastFinishedPulling="2026-03-12 13:30:09.573316884 +0000 UTC m=+1228.022012290" observedRunningTime="2026-03-12 13:30:12.203229771 +0000 UTC m=+1230.651925157" watchObservedRunningTime="2026-03-12 13:30:12.207344289 +0000 UTC m=+1230.656039685" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.226410 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" podStartSLOduration=6.955253243 podStartE2EDuration="2m25.226389281s" podCreationTimestamp="2026-03-12 13:27:47 +0000 UTC" firstStartedPulling="2026-03-12 13:27:51.660989343 +0000 UTC m=+1090.109684749" lastFinishedPulling="2026-03-12 13:30:09.932125351 +0000 UTC m=+1228.380820787" observedRunningTime="2026-03-12 13:30:12.222271114 +0000 UTC m=+1230.670966510" watchObservedRunningTime="2026-03-12 13:30:12.226389281 +0000 UTC m=+1230.675084677" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.257772 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" podStartSLOduration=8.54365231 podStartE2EDuration="2m25.257756855s" podCreationTimestamp="2026-03-12 13:27:47 +0000 UTC" firstStartedPulling="2026-03-12 13:27:52.859319182 +0000 UTC m=+1091.308014578" lastFinishedPulling="2026-03-12 13:30:09.573423717 +0000 UTC m=+1228.022119123" observedRunningTime="2026-03-12 13:30:12.252787864 +0000 UTC m=+1230.701483260" watchObservedRunningTime="2026-03-12 13:30:12.257756855 +0000 UTC m=+1230.706452251" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.305632 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" podStartSLOduration=23.456396545 podStartE2EDuration="2m24.305611239s" podCreationTimestamp="2026-03-12 13:27:48 +0000 UTC" firstStartedPulling="2026-03-12 13:27:51.660382316 +0000 UTC m=+1090.109077712" lastFinishedPulling="2026-03-12 13:29:52.50959699 +0000 UTC m=+1210.958292406" observedRunningTime="2026-03-12 13:30:12.29861577 +0000 UTC m=+1230.747311166" watchObservedRunningTime="2026-03-12 13:30:12.305611239 +0000 UTC m=+1230.754306635" Mar 12 13:30:12 crc kubenswrapper[4778]: I0312 13:30:12.307940 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" podStartSLOduration=7.948900162 podStartE2EDuration="2m24.307927365s" podCreationTimestamp="2026-03-12 13:27:48 +0000 UTC" firstStartedPulling="2026-03-12 13:27:53.997457803 +0000 UTC m=+1092.446153199" lastFinishedPulling="2026-03-12 13:30:10.356485006 +0000 UTC m=+1228.805180402" observedRunningTime="2026-03-12 13:30:12.275092039 +0000 UTC m=+1230.723787435" watchObservedRunningTime="2026-03-12 13:30:12.307927365 +0000 UTC m=+1230.756622771" Mar 12 13:30:13 crc kubenswrapper[4778]: I0312 13:30:13.075209 4778 generic.go:334] "Generic (PLEG): container finished" podID="8bf03685-d980-41f0-bbc5-84b9ae0ce1df" containerID="fa067a709ad1af5d5b9327929891ffc04839dd2d8aba3cc70c48dbfeabd353b9" exitCode=0 Mar 12 13:30:13 crc kubenswrapper[4778]: I0312 13:30:13.075295 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" event={"ID":"8bf03685-d980-41f0-bbc5-84b9ae0ce1df","Type":"ContainerDied","Data":"fa067a709ad1af5d5b9327929891ffc04839dd2d8aba3cc70c48dbfeabd353b9"} Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.312711 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.411379 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xsfh\" (UniqueName: \"kubernetes.io/projected/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-kube-api-access-4xsfh\") pod \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.412548 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-secret-volume\") pod \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.413073 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-config-volume\") pod \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\" (UID: \"8bf03685-d980-41f0-bbc5-84b9ae0ce1df\") " Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.413671 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bf03685-d980-41f0-bbc5-84b9ae0ce1df" (UID: "8bf03685-d980-41f0-bbc5-84b9ae0ce1df"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.414440 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.419528 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-kube-api-access-4xsfh" (OuterVolumeSpecName: "kube-api-access-4xsfh") pod "8bf03685-d980-41f0-bbc5-84b9ae0ce1df" (UID: "8bf03685-d980-41f0-bbc5-84b9ae0ce1df"). InnerVolumeSpecName "kube-api-access-4xsfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.420163 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bf03685-d980-41f0-bbc5-84b9ae0ce1df" (UID: "8bf03685-d980-41f0-bbc5-84b9ae0ce1df"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.516247 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:30:14 crc kubenswrapper[4778]: I0312 13:30:14.516281 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xsfh\" (UniqueName: \"kubernetes.io/projected/8bf03685-d980-41f0-bbc5-84b9ae0ce1df-kube-api-access-4xsfh\") on node \"crc\" DevicePath \"\"" Mar 12 13:30:15 crc kubenswrapper[4778]: I0312 13:30:15.090963 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" event={"ID":"8bf03685-d980-41f0-bbc5-84b9ae0ce1df","Type":"ContainerDied","Data":"1a5e91c9d02c7de2c4009f12029cabe072d8d04d8718617498c0566b1d22e0df"} Mar 12 13:30:15 crc kubenswrapper[4778]: I0312 13:30:15.091049 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a5e91c9d02c7de2c4009f12029cabe072d8d04d8718617498c0566b1d22e0df" Mar 12 13:30:15 crc kubenswrapper[4778]: I0312 13:30:15.091441 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f" Mar 12 13:30:16 crc kubenswrapper[4778]: I0312 13:30:16.099552 4778 generic.go:334] "Generic (PLEG): container finished" podID="1c682acb-240b-44d4-a2be-0ea0cd913af1" containerID="8328194fef169053b3f39722ffd3e2d940869363b5142050b8e768ed01fab0c0" exitCode=0 Mar 12 13:30:16 crc kubenswrapper[4778]: I0312 13:30:16.099617 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" event={"ID":"1c682acb-240b-44d4-a2be-0ea0cd913af1","Type":"ContainerDied","Data":"8328194fef169053b3f39722ffd3e2d940869363b5142050b8e768ed01fab0c0"} Mar 12 13:30:17 crc kubenswrapper[4778]: I0312 13:30:17.386883 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:17 crc kubenswrapper[4778]: I0312 13:30:17.560830 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ks2n\" (UniqueName: \"kubernetes.io/projected/1c682acb-240b-44d4-a2be-0ea0cd913af1-kube-api-access-9ks2n\") pod \"1c682acb-240b-44d4-a2be-0ea0cd913af1\" (UID: \"1c682acb-240b-44d4-a2be-0ea0cd913af1\") " Mar 12 13:30:17 crc kubenswrapper[4778]: I0312 13:30:17.569838 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c682acb-240b-44d4-a2be-0ea0cd913af1-kube-api-access-9ks2n" (OuterVolumeSpecName: "kube-api-access-9ks2n") pod "1c682acb-240b-44d4-a2be-0ea0cd913af1" (UID: "1c682acb-240b-44d4-a2be-0ea0cd913af1"). InnerVolumeSpecName "kube-api-access-9ks2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:30:17 crc kubenswrapper[4778]: I0312 13:30:17.662970 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ks2n\" (UniqueName: \"kubernetes.io/projected/1c682acb-240b-44d4-a2be-0ea0cd913af1-kube-api-access-9ks2n\") on node \"crc\" DevicePath \"\"" Mar 12 13:30:17 crc kubenswrapper[4778]: I0312 13:30:17.695760 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6" Mar 12 13:30:18 crc kubenswrapper[4778]: I0312 13:30:18.120551 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" event={"ID":"1c682acb-240b-44d4-a2be-0ea0cd913af1","Type":"ContainerDied","Data":"623f846d223ee24a6cba599fe831dbd5bf60da2ddff32e7acc90057e22b71876"} Mar 12 13:30:18 crc kubenswrapper[4778]: I0312 13:30:18.121179 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="623f846d223ee24a6cba599fe831dbd5bf60da2ddff32e7acc90057e22b71876" Mar 12 13:30:18 crc kubenswrapper[4778]: I0312 13:30:18.120652 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555370-6zrgd" Mar 12 13:30:18 crc kubenswrapper[4778]: I0312 13:30:18.455565 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555364-hrrdv"] Mar 12 13:30:18 crc kubenswrapper[4778]: I0312 13:30:18.461800 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555364-hrrdv"] Mar 12 13:30:18 crc kubenswrapper[4778]: I0312 13:30:18.530036 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-xm4cc" Mar 12 13:30:19 crc kubenswrapper[4778]: I0312 13:30:19.267607 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-9n6jv" Mar 12 13:30:19 crc kubenswrapper[4778]: I0312 13:30:19.377839 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-b7tkm" Mar 12 13:30:19 crc kubenswrapper[4778]: I0312 13:30:19.747443 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-4jgt8" Mar 12 13:30:19 crc kubenswrapper[4778]: I0312 13:30:19.990898 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-dd2ft" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.030293 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-686d5f9fbd-vv9rc" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.125593 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-qb8s8" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.261724 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c862c78c-5987-48cc-8b41-531755f319e9" path="/var/lib/kubelet/pods/c862c78c-5987-48cc-8b41-531755f319e9/volumes" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.282937 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.283009 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7dxdh" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.376508 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cdgg9" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.878793 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-bbgmb" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.979761 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:30:20 crc kubenswrapper[4778]: I0312 13:30:20.982739 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-wvpf8" Mar 12 13:30:21 crc kubenswrapper[4778]: I0312 13:30:21.091936 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-gfv5z" Mar 12 13:30:21 crc kubenswrapper[4778]: I0312 13:30:21.541175 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2tjsk" Mar 12 13:30:25 crc kubenswrapper[4778]: I0312 13:30:25.167916 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-5d6qz" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.743140 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mcpvm"] Mar 12 13:30:41 crc kubenswrapper[4778]: E0312 13:30:41.744066 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c682acb-240b-44d4-a2be-0ea0cd913af1" containerName="oc" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.744078 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c682acb-240b-44d4-a2be-0ea0cd913af1" containerName="oc" Mar 12 13:30:41 crc kubenswrapper[4778]: E0312 13:30:41.744107 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf03685-d980-41f0-bbc5-84b9ae0ce1df" containerName="collect-profiles" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.744113 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf03685-d980-41f0-bbc5-84b9ae0ce1df" containerName="collect-profiles" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.744363 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf03685-d980-41f0-bbc5-84b9ae0ce1df" containerName="collect-profiles" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.744381 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c682acb-240b-44d4-a2be-0ea0cd913af1" containerName="oc" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.745463 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.753839 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.753862 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4pjqn" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.753963 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.753996 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.763980 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mcpvm"] Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.781028 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fr6p2"] Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.782537 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.785252 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.793435 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fr6p2"] Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.913736 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znmfx\" (UniqueName: \"kubernetes.io/projected/68c74642-7beb-4cb9-86bf-b12beafb4b68-kube-api-access-znmfx\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.913788 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.913821 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-config\") pod \"dnsmasq-dns-675f4bcbfc-mcpvm\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.913840 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-config\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:41 crc kubenswrapper[4778]: I0312 13:30:41.913963 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsbqf\" (UniqueName: \"kubernetes.io/projected/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-kube-api-access-hsbqf\") pod \"dnsmasq-dns-675f4bcbfc-mcpvm\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.015537 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znmfx\" (UniqueName: \"kubernetes.io/projected/68c74642-7beb-4cb9-86bf-b12beafb4b68-kube-api-access-znmfx\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.015627 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.015656 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-config\") pod \"dnsmasq-dns-675f4bcbfc-mcpvm\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.015674 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-config\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.015697 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsbqf\" (UniqueName: \"kubernetes.io/projected/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-kube-api-access-hsbqf\") pod \"dnsmasq-dns-675f4bcbfc-mcpvm\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.016710 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-config\") pod \"dnsmasq-dns-675f4bcbfc-mcpvm\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.016739 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-config\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.016768 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.033833 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsbqf\" (UniqueName: \"kubernetes.io/projected/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-kube-api-access-hsbqf\") pod \"dnsmasq-dns-675f4bcbfc-mcpvm\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.034318 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znmfx\" (UniqueName: \"kubernetes.io/projected/68c74642-7beb-4cb9-86bf-b12beafb4b68-kube-api-access-znmfx\") pod \"dnsmasq-dns-78dd6ddcc-fr6p2\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.070431 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.097546 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.328548 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mcpvm"] Mar 12 13:30:42 crc kubenswrapper[4778]: I0312 13:30:42.364916 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fr6p2"] Mar 12 13:30:42 crc kubenswrapper[4778]: W0312 13:30:42.380034 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68c74642_7beb_4cb9_86bf_b12beafb4b68.slice/crio-f62bb5eb4a941e7684e59f1fde389da442d607aaf8bb4fcbc4a589c4e0b98935 WatchSource:0}: Error finding container f62bb5eb4a941e7684e59f1fde389da442d607aaf8bb4fcbc4a589c4e0b98935: Status 404 returned error can't find the container with id f62bb5eb4a941e7684e59f1fde389da442d607aaf8bb4fcbc4a589c4e0b98935 Mar 12 13:30:43 crc kubenswrapper[4778]: I0312 13:30:43.325996 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" event={"ID":"68c74642-7beb-4cb9-86bf-b12beafb4b68","Type":"ContainerStarted","Data":"f62bb5eb4a941e7684e59f1fde389da442d607aaf8bb4fcbc4a589c4e0b98935"} Mar 12 13:30:43 crc kubenswrapper[4778]: I0312 13:30:43.327016 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" event={"ID":"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c","Type":"ContainerStarted","Data":"f54a79ac4265517d4761f6dc0e556ce29441b7767ec9275aa7bd3cc4d56d57eb"} Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.530628 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mcpvm"] Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.555905 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-d9gsf"] Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.557240 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.568083 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-d9gsf"] Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.651299 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-config\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.651361 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwddk\" (UniqueName: \"kubernetes.io/projected/78c6f209-08e0-4789-be6e-8c319547338c-kube-api-access-hwddk\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.651472 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.757139 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-config\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.757278 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwddk\" (UniqueName: \"kubernetes.io/projected/78c6f209-08e0-4789-be6e-8c319547338c-kube-api-access-hwddk\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.757382 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.758468 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.758551 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-config\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.784466 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwddk\" (UniqueName: \"kubernetes.io/projected/78c6f209-08e0-4789-be6e-8c319547338c-kube-api-access-hwddk\") pod \"dnsmasq-dns-5ccc8479f9-d9gsf\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.862261 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fr6p2"] Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.885370 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2p4pj"] Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.885659 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.891620 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:44 crc kubenswrapper[4778]: I0312 13:30:44.905651 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2p4pj"] Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.062960 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.063046 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-config\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.063214 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcntw\" (UniqueName: \"kubernetes.io/projected/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-kube-api-access-zcntw\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.164093 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.164174 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-config\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.164233 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcntw\" (UniqueName: \"kubernetes.io/projected/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-kube-api-access-zcntw\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.165242 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.165729 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-config\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.188164 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcntw\" (UniqueName: \"kubernetes.io/projected/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-kube-api-access-zcntw\") pod \"dnsmasq-dns-57d769cc4f-2p4pj\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.210909 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.433100 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-d9gsf"] Mar 12 13:30:45 crc kubenswrapper[4778]: W0312 13:30:45.444120 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78c6f209_08e0_4789_be6e_8c319547338c.slice/crio-e3e2879145875855639170cdeac27dda0895e629f9eadf854b4f0adb8048db0a WatchSource:0}: Error finding container e3e2879145875855639170cdeac27dda0895e629f9eadf854b4f0adb8048db0a: Status 404 returned error can't find the container with id e3e2879145875855639170cdeac27dda0895e629f9eadf854b4f0adb8048db0a Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.698739 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2p4pj"] Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.703788 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.706197 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: W0312 13:30:45.706474 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc10b98ea_d832_471e_adb6_c22c4dbb0ab8.slice/crio-0799732deec1f1a8aef551ea0f0b4139ada27fd6ead6c91498a4273deb0bea7d WatchSource:0}: Error finding container 0799732deec1f1a8aef551ea0f0b4139ada27fd6ead6c91498a4273deb0bea7d: Status 404 returned error can't find the container with id 0799732deec1f1a8aef551ea0f0b4139ada27fd6ead6c91498a4273deb0bea7d Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.708118 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.710401 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.710441 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-n2fr8" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.710468 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.710570 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.710645 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.710756 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.714453 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874102 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874142 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874177 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874225 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kcbq\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-kube-api-access-7kcbq\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874255 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874278 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874304 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874327 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874359 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874380 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.874404 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977328 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kcbq\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-kube-api-access-7kcbq\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977483 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977514 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977564 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977599 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977670 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977698 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977751 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977789 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977810 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.977857 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.978013 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.978283 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.978459 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.979773 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.980324 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.980543 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.984937 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.985046 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.994235 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.994319 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:45 crc kubenswrapper[4778]: I0312 13:30:45.997320 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kcbq\" (UniqueName: \"kubernetes.io/projected/629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03-kube-api-access-7kcbq\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.011655 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.013427 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.018594 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.018759 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.018956 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.019093 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.019296 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.019410 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.019505 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.019619 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r9xhc" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.030784 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.037444 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184000 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184116 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184219 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184265 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e89dfcc-2ac3-444c-91e8-56991eae096b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184313 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-config-data\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184385 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e89dfcc-2ac3-444c-91e8-56991eae096b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184410 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184454 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184483 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184531 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.184552 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kxn5\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-kube-api-access-4kxn5\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.285922 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286261 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e89dfcc-2ac3-444c-91e8-56991eae096b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286297 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-config-data\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286322 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e89dfcc-2ac3-444c-91e8-56991eae096b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286343 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286369 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286396 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286416 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286433 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kxn5\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-kube-api-access-4kxn5\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286472 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.286495 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.287386 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.287670 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") device mount path \"/mnt/openstack/pv19\"" pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.299147 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.300410 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-config-data\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.301583 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.302148 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1e89dfcc-2ac3-444c-91e8-56991eae096b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.310279 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.310471 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1e89dfcc-2ac3-444c-91e8-56991eae096b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.316933 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1e89dfcc-2ac3-444c-91e8-56991eae096b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.320917 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.331237 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kxn5\" (UniqueName: \"kubernetes.io/projected/1e89dfcc-2ac3-444c-91e8-56991eae096b-kube-api-access-4kxn5\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.383549 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"rabbitmq-server-0\" (UID: \"1e89dfcc-2ac3-444c-91e8-56991eae096b\") " pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.402511 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" event={"ID":"78c6f209-08e0-4789-be6e-8c319547338c","Type":"ContainerStarted","Data":"e3e2879145875855639170cdeac27dda0895e629f9eadf854b4f0adb8048db0a"} Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.432060 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" event={"ID":"c10b98ea-d832-471e-adb6-c22c4dbb0ab8","Type":"ContainerStarted","Data":"0799732deec1f1a8aef551ea0f0b4139ada27fd6ead6c91498a4273deb0bea7d"} Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.670130 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 13:30:46 crc kubenswrapper[4778]: W0312 13:30:46.694605 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod629c84c5_e6cf_4aa7_ba9a_5a5fe7f53a03.slice/crio-4136242d28b5d751ac159a0788e9e237a903f60a0018c9e2a52d0ac52cd311ca WatchSource:0}: Error finding container 4136242d28b5d751ac159a0788e9e237a903f60a0018c9e2a52d0ac52cd311ca: Status 404 returned error can't find the container with id 4136242d28b5d751ac159a0788e9e237a903f60a0018c9e2a52d0ac52cd311ca Mar 12 13:30:46 crc kubenswrapper[4778]: I0312 13:30:46.695774 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.172661 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.353966 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.355354 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.359654 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.359927 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.360835 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-cj5pw" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.361033 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.365261 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.370955 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.474366 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e89dfcc-2ac3-444c-91e8-56991eae096b","Type":"ContainerStarted","Data":"3c486631d5b69b991065345f5c9738213bf611e7c3f421f730777ed8f23e1701"} Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.477653 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03","Type":"ContainerStarted","Data":"4136242d28b5d751ac159a0788e9e237a903f60a0018c9e2a52d0ac52cd311ca"} Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.517709 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-kolla-config\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518110 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/663feb48-0ed1-4947-97c3-e0bac206fdb2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518167 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518254 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-config-data-default\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518291 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/663feb48-0ed1-4947-97c3-e0bac206fdb2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518318 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9rpt\" (UniqueName: \"kubernetes.io/projected/663feb48-0ed1-4947-97c3-e0bac206fdb2-kube-api-access-k9rpt\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518344 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.518371 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663feb48-0ed1-4947-97c3-e0bac206fdb2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625512 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625588 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-config-data-default\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625616 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/663feb48-0ed1-4947-97c3-e0bac206fdb2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625636 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9rpt\" (UniqueName: \"kubernetes.io/projected/663feb48-0ed1-4947-97c3-e0bac206fdb2-kube-api-access-k9rpt\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625655 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625677 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663feb48-0ed1-4947-97c3-e0bac206fdb2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625710 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-kolla-config\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.625740 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/663feb48-0ed1-4947-97c3-e0bac206fdb2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.627371 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/663feb48-0ed1-4947-97c3-e0bac206fdb2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.628631 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.630163 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-kolla-config\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.630998 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.632997 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/663feb48-0ed1-4947-97c3-e0bac206fdb2-config-data-default\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.640525 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663feb48-0ed1-4947-97c3-e0bac206fdb2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.645889 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/663feb48-0ed1-4947-97c3-e0bac206fdb2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.647434 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9rpt\" (UniqueName: \"kubernetes.io/projected/663feb48-0ed1-4947-97c3-e0bac206fdb2-kube-api-access-k9rpt\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.655772 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"663feb48-0ed1-4947-97c3-e0bac206fdb2\") " pod="openstack/openstack-galera-0" Mar 12 13:30:47 crc kubenswrapper[4778]: I0312 13:30:47.742608 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.330406 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.755760 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.757809 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.759921 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-79jqq" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.760143 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.760390 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.760420 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.792004 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849103 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvkn6\" (UniqueName: \"kubernetes.io/projected/fe52f8ba-9053-4733-b2e3-8f1becf437c8-kube-api-access-hvkn6\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849199 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849230 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe52f8ba-9053-4733-b2e3-8f1becf437c8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849251 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe52f8ba-9053-4733-b2e3-8f1becf437c8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849288 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849319 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849342 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.849357 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe52f8ba-9053-4733-b2e3-8f1becf437c8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.950880 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe52f8ba-9053-4733-b2e3-8f1becf437c8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.950961 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951014 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951048 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951062 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe52f8ba-9053-4733-b2e3-8f1becf437c8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951097 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvkn6\" (UniqueName: \"kubernetes.io/projected/fe52f8ba-9053-4733-b2e3-8f1becf437c8-kube-api-access-hvkn6\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951148 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951551 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe52f8ba-9053-4733-b2e3-8f1becf437c8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.951878 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.952115 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe52f8ba-9053-4733-b2e3-8f1becf437c8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.953325 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.953766 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.956241 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe52f8ba-9053-4733-b2e3-8f1becf437c8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.957347 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe52f8ba-9053-4733-b2e3-8f1becf437c8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.961794 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe52f8ba-9053-4733-b2e3-8f1becf437c8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.976974 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvkn6\" (UniqueName: \"kubernetes.io/projected/fe52f8ba-9053-4733-b2e3-8f1becf437c8-kube-api-access-hvkn6\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:48 crc kubenswrapper[4778]: I0312 13:30:48.978705 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fe52f8ba-9053-4733-b2e3-8f1becf437c8\") " pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.080677 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.091367 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.096597 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.097839 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jrgp8" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.097937 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.101144 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.103589 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.255700 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zj26\" (UniqueName: \"kubernetes.io/projected/ec63cc68-6fde-419b-973c-91fc982e6a49-kube-api-access-6zj26\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.255746 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec63cc68-6fde-419b-973c-91fc982e6a49-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.255811 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec63cc68-6fde-419b-973c-91fc982e6a49-config-data\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.255837 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ec63cc68-6fde-419b-973c-91fc982e6a49-kolla-config\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.255856 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec63cc68-6fde-419b-973c-91fc982e6a49-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.357049 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec63cc68-6fde-419b-973c-91fc982e6a49-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.357141 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec63cc68-6fde-419b-973c-91fc982e6a49-config-data\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.357166 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ec63cc68-6fde-419b-973c-91fc982e6a49-kolla-config\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.357195 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec63cc68-6fde-419b-973c-91fc982e6a49-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.357250 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zj26\" (UniqueName: \"kubernetes.io/projected/ec63cc68-6fde-419b-973c-91fc982e6a49-kube-api-access-6zj26\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.358364 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ec63cc68-6fde-419b-973c-91fc982e6a49-kolla-config\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.358831 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec63cc68-6fde-419b-973c-91fc982e6a49-config-data\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.363135 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec63cc68-6fde-419b-973c-91fc982e6a49-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.376421 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec63cc68-6fde-419b-973c-91fc982e6a49-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.384748 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zj26\" (UniqueName: \"kubernetes.io/projected/ec63cc68-6fde-419b-973c-91fc982e6a49-kube-api-access-6zj26\") pod \"memcached-0\" (UID: \"ec63cc68-6fde-419b-973c-91fc982e6a49\") " pod="openstack/memcached-0" Mar 12 13:30:49 crc kubenswrapper[4778]: I0312 13:30:49.428971 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.270019 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.273613 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.276560 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-skn4q" Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.276700 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.396095 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8s49\" (UniqueName: \"kubernetes.io/projected/66ed2760-88a0-4731-a0d1-52cb6cffa2b1-kube-api-access-m8s49\") pod \"kube-state-metrics-0\" (UID: \"66ed2760-88a0-4731-a0d1-52cb6cffa2b1\") " pod="openstack/kube-state-metrics-0" Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.498457 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8s49\" (UniqueName: \"kubernetes.io/projected/66ed2760-88a0-4731-a0d1-52cb6cffa2b1-kube-api-access-m8s49\") pod \"kube-state-metrics-0\" (UID: \"66ed2760-88a0-4731-a0d1-52cb6cffa2b1\") " pod="openstack/kube-state-metrics-0" Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.530980 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8s49\" (UniqueName: \"kubernetes.io/projected/66ed2760-88a0-4731-a0d1-52cb6cffa2b1-kube-api-access-m8s49\") pod \"kube-state-metrics-0\" (UID: \"66ed2760-88a0-4731-a0d1-52cb6cffa2b1\") " pod="openstack/kube-state-metrics-0" Mar 12 13:30:51 crc kubenswrapper[4778]: I0312 13:30:51.606730 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.421166 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4wct6"] Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.422583 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.425073 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-467ql" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.425093 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.425463 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.436683 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4wct6"] Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.450426 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-p67vh"] Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.452061 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.458268 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-p67vh"] Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.551898 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-log-ovn\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.551943 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8efd1e-884d-4963-b69f-04ede0a92267-scripts\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.551963 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-run\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.551979 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h6ww\" (UniqueName: \"kubernetes.io/projected/3b8efd1e-884d-4963-b69f-04ede0a92267-kube-api-access-7h6ww\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552005 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-etc-ovs\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552037 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-lib\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552060 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-run-ovn\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552096 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-run\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552113 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsnrr\" (UniqueName: \"kubernetes.io/projected/bd159b65-0c66-4809-949e-0f1babbaa8e6-kube-api-access-lsnrr\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552127 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8efd1e-884d-4963-b69f-04ede0a92267-ovn-controller-tls-certs\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552161 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-log\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552178 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd159b65-0c66-4809-949e-0f1babbaa8e6-scripts\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.552210 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8efd1e-884d-4963-b69f-04ede0a92267-combined-ca-bundle\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.568294 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"663feb48-0ed1-4947-97c3-e0bac206fdb2","Type":"ContainerStarted","Data":"a3429a93d8a521bc99b465d949c9e95c96869fd73c2a7bc50e7d8f8ffa485293"} Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654093 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-log\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654168 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd159b65-0c66-4809-949e-0f1babbaa8e6-scripts\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654216 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8efd1e-884d-4963-b69f-04ede0a92267-combined-ca-bundle\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654248 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-log-ovn\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654288 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8efd1e-884d-4963-b69f-04ede0a92267-scripts\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654312 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-run\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654335 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h6ww\" (UniqueName: \"kubernetes.io/projected/3b8efd1e-884d-4963-b69f-04ede0a92267-kube-api-access-7h6ww\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654363 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-etc-ovs\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654401 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-lib\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654432 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-run-ovn\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654477 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-run\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654501 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsnrr\" (UniqueName: \"kubernetes.io/projected/bd159b65-0c66-4809-949e-0f1babbaa8e6-kube-api-access-lsnrr\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.654523 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8efd1e-884d-4963-b69f-04ede0a92267-ovn-controller-tls-certs\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.655804 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-run\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.655941 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-log\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.657983 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd159b65-0c66-4809-949e-0f1babbaa8e6-scripts\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.658402 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-log-ovn\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.658546 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-var-lib\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.658582 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-run-ovn\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.658714 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bd159b65-0c66-4809-949e-0f1babbaa8e6-etc-ovs\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.662357 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8efd1e-884d-4963-b69f-04ede0a92267-combined-ca-bundle\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.662722 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8efd1e-884d-4963-b69f-04ede0a92267-ovn-controller-tls-certs\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.673376 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h6ww\" (UniqueName: \"kubernetes.io/projected/3b8efd1e-884d-4963-b69f-04ede0a92267-kube-api-access-7h6ww\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.675524 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8efd1e-884d-4963-b69f-04ede0a92267-scripts\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.675653 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3b8efd1e-884d-4963-b69f-04ede0a92267-var-run\") pod \"ovn-controller-4wct6\" (UID: \"3b8efd1e-884d-4963-b69f-04ede0a92267\") " pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.677565 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsnrr\" (UniqueName: \"kubernetes.io/projected/bd159b65-0c66-4809-949e-0f1babbaa8e6-kube-api-access-lsnrr\") pod \"ovn-controller-ovs-p67vh\" (UID: \"bd159b65-0c66-4809-949e-0f1babbaa8e6\") " pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.747104 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4wct6" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.767266 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.862576 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.954497 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.956068 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.958497 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.959056 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.959359 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.959444 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-9h9p5" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.959543 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 12 13:30:54 crc kubenswrapper[4778]: I0312 13:30:54.973667 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058288 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058407 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058552 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7321e15e-673c-4e0d-80f8-6ac644c1940f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058590 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058612 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5sss\" (UniqueName: \"kubernetes.io/projected/7321e15e-673c-4e0d-80f8-6ac644c1940f-kube-api-access-j5sss\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058731 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7321e15e-673c-4e0d-80f8-6ac644c1940f-config\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058775 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.058805 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7321e15e-673c-4e0d-80f8-6ac644c1940f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160075 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7321e15e-673c-4e0d-80f8-6ac644c1940f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160172 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160257 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5sss\" (UniqueName: \"kubernetes.io/projected/7321e15e-673c-4e0d-80f8-6ac644c1940f-kube-api-access-j5sss\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160382 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7321e15e-673c-4e0d-80f8-6ac644c1940f-config\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160426 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160459 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7321e15e-673c-4e0d-80f8-6ac644c1940f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160552 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.160610 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.161119 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") device mount path \"/mnt/openstack/pv14\"" pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.161231 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7321e15e-673c-4e0d-80f8-6ac644c1940f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.161447 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7321e15e-673c-4e0d-80f8-6ac644c1940f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.165339 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7321e15e-673c-4e0d-80f8-6ac644c1940f-config\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.168106 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.178734 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.178990 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7321e15e-673c-4e0d-80f8-6ac644c1940f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.181717 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5sss\" (UniqueName: \"kubernetes.io/projected/7321e15e-673c-4e0d-80f8-6ac644c1940f-kube-api-access-j5sss\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.196515 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7321e15e-673c-4e0d-80f8-6ac644c1940f\") " pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:55 crc kubenswrapper[4778]: I0312 13:30:55.272636 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.698414 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.705256 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.707308 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.707763 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.707998 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.708914 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7hqfg" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.711843 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.820841 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.820899 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm2k4\" (UniqueName: \"kubernetes.io/projected/7c951c6f-06fd-4793-a95b-26b5c1400d73-kube-api-access-nm2k4\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.820934 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.820980 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c951c6f-06fd-4793-a95b-26b5c1400d73-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.821021 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c951c6f-06fd-4793-a95b-26b5c1400d73-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.821057 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.821551 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c951c6f-06fd-4793-a95b-26b5c1400d73-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.821647 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923232 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923320 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c951c6f-06fd-4793-a95b-26b5c1400d73-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923341 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923396 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923411 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm2k4\" (UniqueName: \"kubernetes.io/projected/7c951c6f-06fd-4793-a95b-26b5c1400d73-kube-api-access-nm2k4\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923427 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923445 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c951c6f-06fd-4793-a95b-26b5c1400d73-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.923472 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c951c6f-06fd-4793-a95b-26b5c1400d73-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.924071 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c951c6f-06fd-4793-a95b-26b5c1400d73-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.924356 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.925346 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c951c6f-06fd-4793-a95b-26b5c1400d73-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.925414 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c951c6f-06fd-4793-a95b-26b5c1400d73-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.933235 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.934178 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.940047 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c951c6f-06fd-4793-a95b-26b5c1400d73-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.943653 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:58 crc kubenswrapper[4778]: I0312 13:30:58.943828 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm2k4\" (UniqueName: \"kubernetes.io/projected/7c951c6f-06fd-4793-a95b-26b5c1400d73-kube-api-access-nm2k4\") pod \"ovsdbserver-sb-0\" (UID: \"7c951c6f-06fd-4793-a95b-26b5c1400d73\") " pod="openstack/ovsdbserver-sb-0" Mar 12 13:30:59 crc kubenswrapper[4778]: I0312 13:30:59.026405 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 12 13:31:02 crc kubenswrapper[4778]: W0312 13:31:02.838897 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe52f8ba_9053_4733_b2e3_8f1becf437c8.slice/crio-4ed3e2e6288dae34d36e51f67c70e30ae227720d9dfbf94561de11abc987edce WatchSource:0}: Error finding container 4ed3e2e6288dae34d36e51f67c70e30ae227720d9dfbf94561de11abc987edce: Status 404 returned error can't find the container with id 4ed3e2e6288dae34d36e51f67c70e30ae227720d9dfbf94561de11abc987edce Mar 12 13:31:02 crc kubenswrapper[4778]: E0312 13:31:02.858799 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 12 13:31:02 crc kubenswrapper[4778]: E0312 13:31:02.859021 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7kcbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:02 crc kubenswrapper[4778]: E0312 13:31:02.860533 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03" Mar 12 13:31:02 crc kubenswrapper[4778]: E0312 13:31:02.873054 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 12 13:31:02 crc kubenswrapper[4778]: E0312 13:31:02.873424 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4kxn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(1e89dfcc-2ac3-444c-91e8-56991eae096b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:02 crc kubenswrapper[4778]: E0312 13:31:02.874614 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="1e89dfcc-2ac3-444c-91e8-56991eae096b" Mar 12 13:31:03 crc kubenswrapper[4778]: I0312 13:31:03.628515 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fe52f8ba-9053-4733-b2e3-8f1becf437c8","Type":"ContainerStarted","Data":"4ed3e2e6288dae34d36e51f67c70e30ae227720d9dfbf94561de11abc987edce"} Mar 12 13:31:03 crc kubenswrapper[4778]: E0312 13:31:03.630290 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03" Mar 12 13:31:03 crc kubenswrapper[4778]: E0312 13:31:03.638791 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="1e89dfcc-2ac3-444c-91e8-56991eae096b" Mar 12 13:31:07 crc kubenswrapper[4778]: I0312 13:31:07.441108 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 12 13:31:07 crc kubenswrapper[4778]: E0312 13:31:07.908701 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 12 13:31:07 crc kubenswrapper[4778]: E0312 13:31:07.909159 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-znmfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-fr6p2_openstack(68c74642-7beb-4cb9-86bf-b12beafb4b68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:07 crc kubenswrapper[4778]: E0312 13:31:07.910630 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" podUID="68c74642-7beb-4cb9-86bf-b12beafb4b68" Mar 12 13:31:08 crc kubenswrapper[4778]: I0312 13:31:08.302959 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 12 13:31:08 crc kubenswrapper[4778]: I0312 13:31:08.308892 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.148258 4778 scope.go:117] "RemoveContainer" containerID="a8f045f157371374b81f9a3098c61d715d2ce620fdfc3121b5f225672622998f" Mar 12 13:31:09 crc kubenswrapper[4778]: W0312 13:31:09.734444 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7321e15e_673c_4e0d_80f8_6ac644c1940f.slice/crio-f36f09305ca04ca2d01c8a383ad06c9da6a054fd0645f5cc5c51b83742f58fac WatchSource:0}: Error finding container f36f09305ca04ca2d01c8a383ad06c9da6a054fd0645f5cc5c51b83742f58fac: Status 404 returned error can't find the container with id f36f09305ca04ca2d01c8a383ad06c9da6a054fd0645f5cc5c51b83742f58fac Mar 12 13:31:09 crc kubenswrapper[4778]: W0312 13:31:09.751215 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec63cc68_6fde_419b_973c_91fc982e6a49.slice/crio-609ddd8d4d204300e5ea572a906e50c99bca89344b4ff9a98871002711d708b9 WatchSource:0}: Error finding container 609ddd8d4d204300e5ea572a906e50c99bca89344b4ff9a98871002711d708b9: Status 404 returned error can't find the container with id 609ddd8d4d204300e5ea572a906e50c99bca89344b4ff9a98871002711d708b9 Mar 12 13:31:09 crc kubenswrapper[4778]: W0312 13:31:09.752167 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66ed2760_88a0_4731_a0d1_52cb6cffa2b1.slice/crio-4e9e8b87b4e8662cb5ee7f6527d7533b6383b322442ecf5f3470e33d6bb4be86 WatchSource:0}: Error finding container 4e9e8b87b4e8662cb5ee7f6527d7533b6383b322442ecf5f3470e33d6bb4be86: Status 404 returned error can't find the container with id 4e9e8b87b4e8662cb5ee7f6527d7533b6383b322442ecf5f3470e33d6bb4be86 Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.813111 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.813641 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hsbqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-mcpvm_openstack(b217b876-3c50-4d5e-8c5b-40e3f1d95b6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.815237 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" podUID="b217b876-3c50-4d5e-8c5b-40e3f1d95b6c" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.849338 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.849496 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zcntw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-2p4pj_openstack(c10b98ea-d832-471e-adb6-c22c4dbb0ab8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.851397 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" podUID="c10b98ea-d832-471e-adb6-c22c4dbb0ab8" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.890910 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.891059 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwddk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-d9gsf_openstack(78c6f209-08e0-4789-be6e-8c319547338c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:09 crc kubenswrapper[4778]: E0312 13:31:09.892603 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" podUID="78c6f209-08e0-4789-be6e-8c319547338c" Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.905222 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.952031 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-config\") pod \"68c74642-7beb-4cb9-86bf-b12beafb4b68\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.952234 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-dns-svc\") pod \"68c74642-7beb-4cb9-86bf-b12beafb4b68\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.952291 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znmfx\" (UniqueName: \"kubernetes.io/projected/68c74642-7beb-4cb9-86bf-b12beafb4b68-kube-api-access-znmfx\") pod \"68c74642-7beb-4cb9-86bf-b12beafb4b68\" (UID: \"68c74642-7beb-4cb9-86bf-b12beafb4b68\") " Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.953079 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68c74642-7beb-4cb9-86bf-b12beafb4b68" (UID: "68c74642-7beb-4cb9-86bf-b12beafb4b68"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.953065 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-config" (OuterVolumeSpecName: "config") pod "68c74642-7beb-4cb9-86bf-b12beafb4b68" (UID: "68c74642-7beb-4cb9-86bf-b12beafb4b68"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:09 crc kubenswrapper[4778]: I0312 13:31:09.955699 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68c74642-7beb-4cb9-86bf-b12beafb4b68-kube-api-access-znmfx" (OuterVolumeSpecName: "kube-api-access-znmfx") pod "68c74642-7beb-4cb9-86bf-b12beafb4b68" (UID: "68c74642-7beb-4cb9-86bf-b12beafb4b68"). InnerVolumeSpecName "kube-api-access-znmfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.054122 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.054547 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c74642-7beb-4cb9-86bf-b12beafb4b68-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.054562 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znmfx\" (UniqueName: \"kubernetes.io/projected/68c74642-7beb-4cb9-86bf-b12beafb4b68-kube-api-access-znmfx\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.321568 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4wct6"] Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.356387 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-p67vh"] Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.443888 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 12 13:31:10 crc kubenswrapper[4778]: W0312 13:31:10.448814 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c951c6f_06fd_4793_a95b_26b5c1400d73.slice/crio-2c2cc1bf1d75f1a721912ce94ba922071438325eb2db84fb97c7b161e7ac53fa WatchSource:0}: Error finding container 2c2cc1bf1d75f1a721912ce94ba922071438325eb2db84fb97c7b161e7ac53fa: Status 404 returned error can't find the container with id 2c2cc1bf1d75f1a721912ce94ba922071438325eb2db84fb97c7b161e7ac53fa Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.686090 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4wct6" event={"ID":"3b8efd1e-884d-4963-b69f-04ede0a92267","Type":"ContainerStarted","Data":"464be4b7c2eaf4085b25163a77ba143b69ebcc719930ac53a1cbfdbbc77387a5"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.687584 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" event={"ID":"68c74642-7beb-4cb9-86bf-b12beafb4b68","Type":"ContainerDied","Data":"f62bb5eb4a941e7684e59f1fde389da442d607aaf8bb4fcbc4a589c4e0b98935"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.687669 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fr6p2" Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.688910 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c951c6f-06fd-4793-a95b-26b5c1400d73","Type":"ContainerStarted","Data":"2c2cc1bf1d75f1a721912ce94ba922071438325eb2db84fb97c7b161e7ac53fa"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.691399 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7321e15e-673c-4e0d-80f8-6ac644c1940f","Type":"ContainerStarted","Data":"f36f09305ca04ca2d01c8a383ad06c9da6a054fd0645f5cc5c51b83742f58fac"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.692469 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66ed2760-88a0-4731-a0d1-52cb6cffa2b1","Type":"ContainerStarted","Data":"4e9e8b87b4e8662cb5ee7f6527d7533b6383b322442ecf5f3470e33d6bb4be86"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.695239 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"663feb48-0ed1-4947-97c3-e0bac206fdb2","Type":"ContainerStarted","Data":"73048822b2ee0f9b1c1f7f4661f73503814f16f141df2b6300c8112edc68f8fa"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.697627 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fe52f8ba-9053-4733-b2e3-8f1becf437c8","Type":"ContainerStarted","Data":"c915e53efe3d1f80e839ccda66e0bb16e04555c81f0a30ce27190c773550f885"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.700977 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p67vh" event={"ID":"bd159b65-0c66-4809-949e-0f1babbaa8e6","Type":"ContainerStarted","Data":"f3170f6c74a1959c49c161a18d30bea16a54ddcbcbf5342404e2b1ea295b59dd"} Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.702231 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ec63cc68-6fde-419b-973c-91fc982e6a49","Type":"ContainerStarted","Data":"609ddd8d4d204300e5ea572a906e50c99bca89344b4ff9a98871002711d708b9"} Mar 12 13:31:10 crc kubenswrapper[4778]: E0312 13:31:10.703706 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" podUID="c10b98ea-d832-471e-adb6-c22c4dbb0ab8" Mar 12 13:31:10 crc kubenswrapper[4778]: E0312 13:31:10.703973 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" podUID="78c6f209-08e0-4789-be6e-8c319547338c" Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.739238 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fr6p2"] Mar 12 13:31:10 crc kubenswrapper[4778]: I0312 13:31:10.762329 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fr6p2"] Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.340519 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.389377 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-config\") pod \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.389452 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsbqf\" (UniqueName: \"kubernetes.io/projected/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-kube-api-access-hsbqf\") pod \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\" (UID: \"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c\") " Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.390300 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-config" (OuterVolumeSpecName: "config") pod "b217b876-3c50-4d5e-8c5b-40e3f1d95b6c" (UID: "b217b876-3c50-4d5e-8c5b-40e3f1d95b6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.401485 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-kube-api-access-hsbqf" (OuterVolumeSpecName: "kube-api-access-hsbqf") pod "b217b876-3c50-4d5e-8c5b-40e3f1d95b6c" (UID: "b217b876-3c50-4d5e-8c5b-40e3f1d95b6c"). InnerVolumeSpecName "kube-api-access-hsbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.491489 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsbqf\" (UniqueName: \"kubernetes.io/projected/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-kube-api-access-hsbqf\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.491522 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.710441 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" event={"ID":"b217b876-3c50-4d5e-8c5b-40e3f1d95b6c","Type":"ContainerDied","Data":"f54a79ac4265517d4761f6dc0e556ce29441b7767ec9275aa7bd3cc4d56d57eb"} Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.710510 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mcpvm" Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.780462 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mcpvm"] Mar 12 13:31:11 crc kubenswrapper[4778]: I0312 13:31:11.790360 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mcpvm"] Mar 12 13:31:12 crc kubenswrapper[4778]: I0312 13:31:12.263772 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68c74642-7beb-4cb9-86bf-b12beafb4b68" path="/var/lib/kubelet/pods/68c74642-7beb-4cb9-86bf-b12beafb4b68/volumes" Mar 12 13:31:12 crc kubenswrapper[4778]: I0312 13:31:12.264387 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b217b876-3c50-4d5e-8c5b-40e3f1d95b6c" path="/var/lib/kubelet/pods/b217b876-3c50-4d5e-8c5b-40e3f1d95b6c/volumes" Mar 12 13:31:13 crc kubenswrapper[4778]: I0312 13:31:13.730787 4778 generic.go:334] "Generic (PLEG): container finished" podID="663feb48-0ed1-4947-97c3-e0bac206fdb2" containerID="73048822b2ee0f9b1c1f7f4661f73503814f16f141df2b6300c8112edc68f8fa" exitCode=0 Mar 12 13:31:13 crc kubenswrapper[4778]: I0312 13:31:13.730914 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"663feb48-0ed1-4947-97c3-e0bac206fdb2","Type":"ContainerDied","Data":"73048822b2ee0f9b1c1f7f4661f73503814f16f141df2b6300c8112edc68f8fa"} Mar 12 13:31:13 crc kubenswrapper[4778]: I0312 13:31:13.733993 4778 generic.go:334] "Generic (PLEG): container finished" podID="fe52f8ba-9053-4733-b2e3-8f1becf437c8" containerID="c915e53efe3d1f80e839ccda66e0bb16e04555c81f0a30ce27190c773550f885" exitCode=0 Mar 12 13:31:13 crc kubenswrapper[4778]: I0312 13:31:13.734055 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fe52f8ba-9053-4733-b2e3-8f1becf437c8","Type":"ContainerDied","Data":"c915e53efe3d1f80e839ccda66e0bb16e04555c81f0a30ce27190c773550f885"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.755970 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fe52f8ba-9053-4733-b2e3-8f1becf437c8","Type":"ContainerStarted","Data":"6b923e83418fc8dd0c8a9c1863eed3de992759ea4fff0ea9e932755b38e6a24f"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.758815 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c951c6f-06fd-4793-a95b-26b5c1400d73","Type":"ContainerStarted","Data":"03daa5981785dae12ef34bc2860b3ebbb499a62f1e5266d3cbe2c29b9bf0010a"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.763881 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7321e15e-673c-4e0d-80f8-6ac644c1940f","Type":"ContainerStarted","Data":"6d8fb734a26f01a2db24b32fcf641a12cbbff7e4c8e2774e61faa1124eea8858"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.770669 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66ed2760-88a0-4731-a0d1-52cb6cffa2b1","Type":"ContainerStarted","Data":"6addcbc9f6e1bd0c36c2127749a9343943bce9503688868083bfb8596a8eda94"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.770800 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.774359 4778 generic.go:334] "Generic (PLEG): container finished" podID="bd159b65-0c66-4809-949e-0f1babbaa8e6" containerID="3d94d19b11275c13d335ced4fca61c11564e4abcc6e74ebdb428d3364a8ab591" exitCode=0 Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.774430 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p67vh" event={"ID":"bd159b65-0c66-4809-949e-0f1babbaa8e6","Type":"ContainerDied","Data":"3d94d19b11275c13d335ced4fca61c11564e4abcc6e74ebdb428d3364a8ab591"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.793822 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ec63cc68-6fde-419b-973c-91fc982e6a49","Type":"ContainerStarted","Data":"cfc94a4126c33c3c862f26c73b0369926b16397e57b29a1116690c8f54a89d03"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.794251 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.824070 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"663feb48-0ed1-4947-97c3-e0bac206fdb2","Type":"ContainerStarted","Data":"c75e427a712dc07d0e96cd91a97f38ade2a32cd7a1f6258673afbbaa743ba85d"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.840836 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.789657983 podStartE2EDuration="28.840818621s" podCreationTimestamp="2026-03-12 13:30:47 +0000 UTC" firstStartedPulling="2026-03-12 13:31:02.842499956 +0000 UTC m=+1281.291195362" lastFinishedPulling="2026-03-12 13:31:09.893660614 +0000 UTC m=+1288.342356000" observedRunningTime="2026-03-12 13:31:15.787038202 +0000 UTC m=+1294.235733598" watchObservedRunningTime="2026-03-12 13:31:15.840818621 +0000 UTC m=+1294.289514017" Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.843215 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4wct6" event={"ID":"3b8efd1e-884d-4963-b69f-04ede0a92267","Type":"ContainerStarted","Data":"891e4089d91ffd634099bdf5e4625738b33178ff8f1fcc2a63c8701621639047"} Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.843365 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4wct6" Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.877335 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.255888923 podStartE2EDuration="24.877313499s" podCreationTimestamp="2026-03-12 13:30:51 +0000 UTC" firstStartedPulling="2026-03-12 13:31:09.788524144 +0000 UTC m=+1288.237219540" lastFinishedPulling="2026-03-12 13:31:15.40994872 +0000 UTC m=+1293.858644116" observedRunningTime="2026-03-12 13:31:15.872612865 +0000 UTC m=+1294.321308261" watchObservedRunningTime="2026-03-12 13:31:15.877313499 +0000 UTC m=+1294.326008885" Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.909011 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.611291679 podStartE2EDuration="29.908992979s" podCreationTimestamp="2026-03-12 13:30:46 +0000 UTC" firstStartedPulling="2026-03-12 13:30:53.582983245 +0000 UTC m=+1272.031678641" lastFinishedPulling="2026-03-12 13:31:09.880684545 +0000 UTC m=+1288.329379941" observedRunningTime="2026-03-12 13:31:15.906339914 +0000 UTC m=+1294.355035320" watchObservedRunningTime="2026-03-12 13:31:15.908992979 +0000 UTC m=+1294.357688375" Mar 12 13:31:15 crc kubenswrapper[4778]: I0312 13:31:15.999047 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.212567884 podStartE2EDuration="26.999022229s" podCreationTimestamp="2026-03-12 13:30:49 +0000 UTC" firstStartedPulling="2026-03-12 13:31:09.780687341 +0000 UTC m=+1288.229382737" lastFinishedPulling="2026-03-12 13:31:14.567141686 +0000 UTC m=+1293.015837082" observedRunningTime="2026-03-12 13:31:15.942571214 +0000 UTC m=+1294.391266620" watchObservedRunningTime="2026-03-12 13:31:15.999022229 +0000 UTC m=+1294.447717645" Mar 12 13:31:16 crc kubenswrapper[4778]: I0312 13:31:16.001311 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4wct6" podStartSLOduration=17.711347796 podStartE2EDuration="22.001301244s" podCreationTimestamp="2026-03-12 13:30:54 +0000 UTC" firstStartedPulling="2026-03-12 13:31:10.343672048 +0000 UTC m=+1288.792367444" lastFinishedPulling="2026-03-12 13:31:14.633625506 +0000 UTC m=+1293.082320892" observedRunningTime="2026-03-12 13:31:15.975623354 +0000 UTC m=+1294.424318760" watchObservedRunningTime="2026-03-12 13:31:16.001301244 +0000 UTC m=+1294.449996650" Mar 12 13:31:16 crc kubenswrapper[4778]: I0312 13:31:16.848245 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p67vh" event={"ID":"bd159b65-0c66-4809-949e-0f1babbaa8e6","Type":"ContainerStarted","Data":"66bdc50de15bbeb963b21f1399497b80530e6a48047c84bd0860d72204092943"} Mar 12 13:31:16 crc kubenswrapper[4778]: I0312 13:31:16.848594 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p67vh" event={"ID":"bd159b65-0c66-4809-949e-0f1babbaa8e6","Type":"ContainerStarted","Data":"ddac7e774dda104b000b6e9560b67447ae0e9b37ef71be096f71fee4dc4966de"} Mar 12 13:31:16 crc kubenswrapper[4778]: E0312 13:31:16.920948 4778 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.32:41514->38.129.56.32:35979: write tcp 38.129.56.32:41514->38.129.56.32:35979: write: broken pipe Mar 12 13:31:17 crc kubenswrapper[4778]: I0312 13:31:17.744089 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 12 13:31:17 crc kubenswrapper[4778]: I0312 13:31:17.744497 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 12 13:31:17 crc kubenswrapper[4778]: I0312 13:31:17.859770 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e89dfcc-2ac3-444c-91e8-56991eae096b","Type":"ContainerStarted","Data":"491cf83ea2b0803c619e4110e5a18dd9c9b6e2cc2bfd596357f59a6a18312dee"} Mar 12 13:31:17 crc kubenswrapper[4778]: I0312 13:31:17.859835 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:31:17 crc kubenswrapper[4778]: I0312 13:31:17.859872 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:31:17 crc kubenswrapper[4778]: I0312 13:31:17.893868 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-p67vh" podStartSLOduration=19.618817742 podStartE2EDuration="23.893848435s" podCreationTimestamp="2026-03-12 13:30:54 +0000 UTC" firstStartedPulling="2026-03-12 13:31:10.360497447 +0000 UTC m=+1288.809192843" lastFinishedPulling="2026-03-12 13:31:14.63552814 +0000 UTC m=+1293.084223536" observedRunningTime="2026-03-12 13:31:16.869882221 +0000 UTC m=+1295.318577637" watchObservedRunningTime="2026-03-12 13:31:17.893848435 +0000 UTC m=+1296.342543831" Mar 12 13:31:18 crc kubenswrapper[4778]: I0312 13:31:18.868477 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7321e15e-673c-4e0d-80f8-6ac644c1940f","Type":"ContainerStarted","Data":"9d1bd71357006dd049f38bd4772731fe6910d7e7f99da405a22aae4c83d47a42"} Mar 12 13:31:18 crc kubenswrapper[4778]: I0312 13:31:18.872923 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c951c6f-06fd-4793-a95b-26b5c1400d73","Type":"ContainerStarted","Data":"efc582c388e483fa01b639f70125dc0c94e861db5b78afcdb00e31bf1cc42a61"} Mar 12 13:31:18 crc kubenswrapper[4778]: I0312 13:31:18.893726 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.142606981 podStartE2EDuration="25.893700614s" podCreationTimestamp="2026-03-12 13:30:53 +0000 UTC" firstStartedPulling="2026-03-12 13:31:09.737676338 +0000 UTC m=+1288.186371734" lastFinishedPulling="2026-03-12 13:31:18.488769971 +0000 UTC m=+1296.937465367" observedRunningTime="2026-03-12 13:31:18.887267251 +0000 UTC m=+1297.335962677" watchObservedRunningTime="2026-03-12 13:31:18.893700614 +0000 UTC m=+1297.342396020" Mar 12 13:31:18 crc kubenswrapper[4778]: I0312 13:31:18.911748 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.887111491 podStartE2EDuration="21.911726467s" podCreationTimestamp="2026-03-12 13:30:57 +0000 UTC" firstStartedPulling="2026-03-12 13:31:10.451872145 +0000 UTC m=+1288.900567541" lastFinishedPulling="2026-03-12 13:31:18.476487131 +0000 UTC m=+1296.925182517" observedRunningTime="2026-03-12 13:31:18.909677448 +0000 UTC m=+1297.358372834" watchObservedRunningTime="2026-03-12 13:31:18.911726467 +0000 UTC m=+1297.360421873" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.026573 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.097365 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.097441 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.272944 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.324320 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.431423 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.880053 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 12 13:31:19 crc kubenswrapper[4778]: I0312 13:31:19.932896 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.026574 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.164403 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.204797 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2p4pj"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.246948 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vtt4z"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.248241 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.252348 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.278390 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kzfk7"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.279855 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vtt4z"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.279978 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.282468 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.287738 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kzfk7"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355068 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-config\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355458 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-combined-ca-bundle\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355527 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-ovn-rundir\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355569 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjgm\" (UniqueName: \"kubernetes.io/projected/199c7ab7-ef93-4b96-a76c-2476f21795ae-kube-api-access-jmjgm\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355644 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355672 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdxh5\" (UniqueName: \"kubernetes.io/projected/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-kube-api-access-jdxh5\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355701 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355727 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-config\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355783 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-ovs-rundir\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.355809 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.388179 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458024 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-ovn-rundir\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458117 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmjgm\" (UniqueName: \"kubernetes.io/projected/199c7ab7-ef93-4b96-a76c-2476f21795ae-kube-api-access-jmjgm\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458212 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458247 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdxh5\" (UniqueName: \"kubernetes.io/projected/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-kube-api-access-jdxh5\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458310 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458342 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-config\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458499 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-ovs-rundir\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458564 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458686 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-config\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458706 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-ovn-rundir\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.458737 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-combined-ca-bundle\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.459141 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-ovs-rundir\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.459180 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.459666 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.459900 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-config\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.459931 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-config\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.468803 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.468837 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-combined-ca-bundle\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.479028 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdxh5\" (UniqueName: \"kubernetes.io/projected/a8484e5d-6f77-407c-81db-0d9b2a6b37fd-kube-api-access-jdxh5\") pod \"ovn-controller-metrics-vtt4z\" (UID: \"a8484e5d-6f77-407c-81db-0d9b2a6b37fd\") " pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.479531 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmjgm\" (UniqueName: \"kubernetes.io/projected/199c7ab7-ef93-4b96-a76c-2476f21795ae-kube-api-access-jmjgm\") pod \"dnsmasq-dns-7fd796d7df-kzfk7\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.489724 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.558247 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-d9gsf"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.579710 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jsqnb"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.582105 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vtt4z" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.584087 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.588015 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.623310 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.638704 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jsqnb"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.661758 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5589g\" (UniqueName: \"kubernetes.io/projected/124fc095-41fd-4e2d-86a1-0aada5c7447f-kube-api-access-5589g\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.661840 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.661894 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-config\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.662044 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.662098 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.740035 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.756432 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-886c-account-create-update-c7kqb"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.759748 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.762712 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.763379 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.763447 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.763483 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5589g\" (UniqueName: \"kubernetes.io/projected/124fc095-41fd-4e2d-86a1-0aada5c7447f-kube-api-access-5589g\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.763525 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.763566 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-config\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.764619 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.764642 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.764922 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.765767 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-config\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.765819 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-886c-account-create-update-c7kqb"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.817038 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5589g\" (UniqueName: \"kubernetes.io/projected/124fc095-41fd-4e2d-86a1-0aada5c7447f-kube-api-access-5589g\") pod \"dnsmasq-dns-86db49b7ff-jsqnb\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.820975 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-gccjh"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.821980 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gccjh" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.829343 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gccjh"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.865792 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-config\") pod \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.865953 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcntw\" (UniqueName: \"kubernetes.io/projected/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-kube-api-access-zcntw\") pod \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.866012 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-dns-svc\") pod \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\" (UID: \"c10b98ea-d832-471e-adb6-c22c4dbb0ab8\") " Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.866162 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfr4k\" (UniqueName: \"kubernetes.io/projected/fc051b32-4b28-4011-9a00-49caa730f074-kube-api-access-wfr4k\") pod \"placement-db-create-gccjh\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " pod="openstack/placement-db-create-gccjh" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.866279 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc051b32-4b28-4011-9a00-49caa730f074-operator-scripts\") pod \"placement-db-create-gccjh\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " pod="openstack/placement-db-create-gccjh" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.866304 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjpbn\" (UniqueName: \"kubernetes.io/projected/7b329f80-bb88-4c5c-91eb-24394cdcc492-kube-api-access-bjpbn\") pod \"placement-886c-account-create-update-c7kqb\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.866342 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b329f80-bb88-4c5c-91eb-24394cdcc492-operator-scripts\") pod \"placement-886c-account-create-update-c7kqb\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.867177 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-config" (OuterVolumeSpecName: "config") pod "c10b98ea-d832-471e-adb6-c22c4dbb0ab8" (UID: "c10b98ea-d832-471e-adb6-c22c4dbb0ab8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.867229 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c10b98ea-d832-471e-adb6-c22c4dbb0ab8" (UID: "c10b98ea-d832-471e-adb6-c22c4dbb0ab8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.871817 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-kube-api-access-zcntw" (OuterVolumeSpecName: "kube-api-access-zcntw") pod "c10b98ea-d832-471e-adb6-c22c4dbb0ab8" (UID: "c10b98ea-d832-471e-adb6-c22c4dbb0ab8"). InnerVolumeSpecName "kube-api-access-zcntw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.887630 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" event={"ID":"c10b98ea-d832-471e-adb6-c22c4dbb0ab8","Type":"ContainerDied","Data":"0799732deec1f1a8aef551ea0f0b4139ada27fd6ead6c91498a4273deb0bea7d"} Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.887719 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2p4pj" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.894375 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03","Type":"ContainerStarted","Data":"4fe9b07cd1599e91138683ca30e9da84b4bd93250ce15e01fd43967606252649"} Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.944146 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.949265 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970551 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc051b32-4b28-4011-9a00-49caa730f074-operator-scripts\") pod \"placement-db-create-gccjh\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " pod="openstack/placement-db-create-gccjh" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970616 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjpbn\" (UniqueName: \"kubernetes.io/projected/7b329f80-bb88-4c5c-91eb-24394cdcc492-kube-api-access-bjpbn\") pod \"placement-886c-account-create-update-c7kqb\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970728 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b329f80-bb88-4c5c-91eb-24394cdcc492-operator-scripts\") pod \"placement-886c-account-create-update-c7kqb\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970790 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfr4k\" (UniqueName: \"kubernetes.io/projected/fc051b32-4b28-4011-9a00-49caa730f074-kube-api-access-wfr4k\") pod \"placement-db-create-gccjh\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " pod="openstack/placement-db-create-gccjh" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970894 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2p4pj"] Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970917 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcntw\" (UniqueName: \"kubernetes.io/projected/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-kube-api-access-zcntw\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970934 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.970975 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c10b98ea-d832-471e-adb6-c22c4dbb0ab8-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.973531 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc051b32-4b28-4011-9a00-49caa730f074-operator-scripts\") pod \"placement-db-create-gccjh\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " pod="openstack/placement-db-create-gccjh" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.981402 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b329f80-bb88-4c5c-91eb-24394cdcc492-operator-scripts\") pod \"placement-886c-account-create-update-c7kqb\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.993003 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjpbn\" (UniqueName: \"kubernetes.io/projected/7b329f80-bb88-4c5c-91eb-24394cdcc492-kube-api-access-bjpbn\") pod \"placement-886c-account-create-update-c7kqb\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:20 crc kubenswrapper[4778]: I0312 13:31:20.996901 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.001732 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfr4k\" (UniqueName: \"kubernetes.io/projected/fc051b32-4b28-4011-9a00-49caa730f074-kube-api-access-wfr4k\") pod \"placement-db-create-gccjh\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " pod="openstack/placement-db-create-gccjh" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.018127 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2p4pj"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.072875 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-dns-svc\") pod \"78c6f209-08e0-4789-be6e-8c319547338c\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.072947 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwddk\" (UniqueName: \"kubernetes.io/projected/78c6f209-08e0-4789-be6e-8c319547338c-kube-api-access-hwddk\") pod \"78c6f209-08e0-4789-be6e-8c319547338c\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.073141 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-config\") pod \"78c6f209-08e0-4789-be6e-8c319547338c\" (UID: \"78c6f209-08e0-4789-be6e-8c319547338c\") " Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.074392 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-config" (OuterVolumeSpecName: "config") pod "78c6f209-08e0-4789-be6e-8c319547338c" (UID: "78c6f209-08e0-4789-be6e-8c319547338c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.076177 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78c6f209-08e0-4789-be6e-8c319547338c" (UID: "78c6f209-08e0-4789-be6e-8c319547338c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.081420 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c6f209-08e0-4789-be6e-8c319547338c-kube-api-access-hwddk" (OuterVolumeSpecName: "kube-api-access-hwddk") pod "78c6f209-08e0-4789-be6e-8c319547338c" (UID: "78c6f209-08e0-4789-be6e-8c319547338c"). InnerVolumeSpecName "kube-api-access-hwddk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.106652 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.108046 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.112615 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.112797 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.112942 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.113447 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5tksr" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.133824 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.161574 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175167 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b25f9c9-784a-4a52-9bb3-02c6c4592702-config\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175287 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b25f9c9-784a-4a52-9bb3-02c6c4592702-scripts\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175343 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175373 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175450 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b25f9c9-784a-4a52-9bb3-02c6c4592702-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175488 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rptgc\" (UniqueName: \"kubernetes.io/projected/1b25f9c9-784a-4a52-9bb3-02c6c4592702-kube-api-access-rptgc\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175526 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175584 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175608 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78c6f209-08e0-4789-be6e-8c319547338c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.175622 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwddk\" (UniqueName: \"kubernetes.io/projected/78c6f209-08e0-4789-be6e-8c319547338c-kube-api-access-hwddk\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.181666 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gccjh" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.250655 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kzfk7"] Mar 12 13:31:21 crc kubenswrapper[4778]: W0312 13:31:21.264376 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod199c7ab7_ef93_4b96_a76c_2476f21795ae.slice/crio-5f1e3433b2aa51609fb7a612d3a57e0e4f8b5c1392d5055466098db076137d25 WatchSource:0}: Error finding container 5f1e3433b2aa51609fb7a612d3a57e0e4f8b5c1392d5055466098db076137d25: Status 404 returned error can't find the container with id 5f1e3433b2aa51609fb7a612d3a57e0e4f8b5c1392d5055466098db076137d25 Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277724 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b25f9c9-784a-4a52-9bb3-02c6c4592702-config\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277785 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b25f9c9-784a-4a52-9bb3-02c6c4592702-scripts\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277831 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277861 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277916 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b25f9c9-784a-4a52-9bb3-02c6c4592702-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277948 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rptgc\" (UniqueName: \"kubernetes.io/projected/1b25f9c9-784a-4a52-9bb3-02c6c4592702-kube-api-access-rptgc\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.277985 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.280049 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b25f9c9-784a-4a52-9bb3-02c6c4592702-config\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.281840 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b25f9c9-784a-4a52-9bb3-02c6c4592702-scripts\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.282460 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.282471 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.283094 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b25f9c9-784a-4a52-9bb3-02c6c4592702-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.288436 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b25f9c9-784a-4a52-9bb3-02c6c4592702-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.303939 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rptgc\" (UniqueName: \"kubernetes.io/projected/1b25f9c9-784a-4a52-9bb3-02c6c4592702-kube-api-access-rptgc\") pod \"ovn-northd-0\" (UID: \"1b25f9c9-784a-4a52-9bb3-02c6c4592702\") " pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.389547 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vtt4z"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.426948 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-886c-account-create-update-c7kqb"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.434733 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.542650 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jsqnb"] Mar 12 13:31:21 crc kubenswrapper[4778]: W0312 13:31:21.581348 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod124fc095_41fd_4e2d_86a1_0aada5c7447f.slice/crio-a6bf49931f5dff6309127ba2cd8c8cdab17c7cf48d3d19393ec4e377e3b30b19 WatchSource:0}: Error finding container a6bf49931f5dff6309127ba2cd8c8cdab17c7cf48d3d19393ec4e377e3b30b19: Status 404 returned error can't find the container with id a6bf49931f5dff6309127ba2cd8c8cdab17c7cf48d3d19393ec4e377e3b30b19 Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.618547 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.631411 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gccjh"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.734910 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kzfk7"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.774144 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-8rkss"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.782127 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8rkss"] Mar 12 13:31:21 crc kubenswrapper[4778]: I0312 13:31:21.782247 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.005619 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vtt4z" event={"ID":"a8484e5d-6f77-407c-81db-0d9b2a6b37fd","Type":"ContainerStarted","Data":"3d9fbf21819b1d517e03213dbd3933cd631f8d1c8b563683a6040fea42ca66f5"} Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.009257 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gccjh" event={"ID":"fc051b32-4b28-4011-9a00-49caa730f074","Type":"ContainerStarted","Data":"58e88409e0c9f35402bf7f7052b8bfab0894113d1308d624a95e9e78a692921f"} Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.011366 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" event={"ID":"199c7ab7-ef93-4b96-a76c-2476f21795ae","Type":"ContainerStarted","Data":"5f1e3433b2aa51609fb7a612d3a57e0e4f8b5c1392d5055466098db076137d25"} Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.014888 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" event={"ID":"124fc095-41fd-4e2d-86a1-0aada5c7447f","Type":"ContainerStarted","Data":"a6bf49931f5dff6309127ba2cd8c8cdab17c7cf48d3d19393ec4e377e3b30b19"} Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.016897 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-886c-account-create-update-c7kqb" event={"ID":"7b329f80-bb88-4c5c-91eb-24394cdcc492","Type":"ContainerStarted","Data":"4931bee01e54e487e69bca458cb118194d8730ef5cb3f2125d81402dd58d404b"} Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.021880 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.021873 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-d9gsf" event={"ID":"78c6f209-08e0-4789-be6e-8c319547338c","Type":"ContainerDied","Data":"e3e2879145875855639170cdeac27dda0895e629f9eadf854b4f0adb8048db0a"} Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.022752 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.090270 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r26t5\" (UniqueName: \"kubernetes.io/projected/5dd405d8-c82b-49d0-a871-1c7c847638df-kube-api-access-r26t5\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.090690 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.090730 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-dns-svc\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.090775 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.090866 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-config\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.144458 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-d9gsf"] Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.153661 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.160751 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-d9gsf"] Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.192276 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.192333 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-dns-svc\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.192409 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.192544 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-config\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.192616 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r26t5\" (UniqueName: \"kubernetes.io/projected/5dd405d8-c82b-49d0-a871-1c7c847638df-kube-api-access-r26t5\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.195336 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.196022 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-dns-svc\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.202359 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-config\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.202472 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.216540 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.221763 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r26t5\" (UniqueName: \"kubernetes.io/projected/5dd405d8-c82b-49d0-a871-1c7c847638df-kube-api-access-r26t5\") pod \"dnsmasq-dns-698758b865-8rkss\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.274651 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c6f209-08e0-4789-be6e-8c319547338c" path="/var/lib/kubelet/pods/78c6f209-08e0-4789-be6e-8c319547338c/volumes" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.276580 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c10b98ea-d832-471e-adb6-c22c4dbb0ab8" path="/var/lib/kubelet/pods/c10b98ea-d832-471e-adb6-c22c4dbb0ab8/volumes" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.362163 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:22 crc kubenswrapper[4778]: I0312 13:31:22.870478 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8rkss"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.034131 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b25f9c9-784a-4a52-9bb3-02c6c4592702","Type":"ContainerStarted","Data":"6b6d08caea52a77ace32f1739d7850b5128d6bc8c08fd6572fc7842be0f72e21"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.044698 4778 generic.go:334] "Generic (PLEG): container finished" podID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerID="2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9" exitCode=0 Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.044798 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" event={"ID":"124fc095-41fd-4e2d-86a1-0aada5c7447f","Type":"ContainerDied","Data":"2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.048222 4778 generic.go:334] "Generic (PLEG): container finished" podID="7b329f80-bb88-4c5c-91eb-24394cdcc492" containerID="2c45c4ddf823adba305999f51111b5e3abaff88105a2366fb93304b13b53f40d" exitCode=0 Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.048411 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-886c-account-create-update-c7kqb" event={"ID":"7b329f80-bb88-4c5c-91eb-24394cdcc492","Type":"ContainerDied","Data":"2c45c4ddf823adba305999f51111b5e3abaff88105a2366fb93304b13b53f40d"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.053003 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8rkss" event={"ID":"5dd405d8-c82b-49d0-a871-1c7c847638df","Type":"ContainerStarted","Data":"0cf55f4c77e0e83cbfd4fa4c9d04d1940beb400c64b78ffec689c21b7bd18ebf"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.054338 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vtt4z" event={"ID":"a8484e5d-6f77-407c-81db-0d9b2a6b37fd","Type":"ContainerStarted","Data":"53eaf5195f1cd4ce28bf1a0b3cd6d8b74553ecbdd4b4f588c17d138d4544e87a"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.058463 4778 generic.go:334] "Generic (PLEG): container finished" podID="fc051b32-4b28-4011-9a00-49caa730f074" containerID="d8bdc9c2c4e5e8d5384ac13e3814d6ad0bf996923ba03462051d4c078107d461" exitCode=0 Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.058618 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gccjh" event={"ID":"fc051b32-4b28-4011-9a00-49caa730f074","Type":"ContainerDied","Data":"d8bdc9c2c4e5e8d5384ac13e3814d6ad0bf996923ba03462051d4c078107d461"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.060398 4778 generic.go:334] "Generic (PLEG): container finished" podID="199c7ab7-ef93-4b96-a76c-2476f21795ae" containerID="4d8c29e71c21a3ab92dc60d0b4d6da588df8f7c85dce7dcda5be210ad42005de" exitCode=0 Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.061413 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" event={"ID":"199c7ab7-ef93-4b96-a76c-2476f21795ae","Type":"ContainerDied","Data":"4d8c29e71c21a3ab92dc60d0b4d6da588df8f7c85dce7dcda5be210ad42005de"} Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.119379 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.136678 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.136816 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.142394 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.145865 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-kfb99" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.151544 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.153637 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.179319 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vtt4z" podStartSLOduration=3.179292738 podStartE2EDuration="3.179292738s" podCreationTimestamp="2026-03-12 13:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:23.158683042 +0000 UTC m=+1301.607378428" watchObservedRunningTime="2026-03-12 13:31:23.179292738 +0000 UTC m=+1301.627988134" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.313061 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01f943c-e09c-4727-8cf7-eec58a56b363-lock\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.313178 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01f943c-e09c-4727-8cf7-eec58a56b363-cache\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.313220 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.313246 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.313266 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01f943c-e09c-4727-8cf7-eec58a56b363-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.313524 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xltnr\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-kube-api-access-xltnr\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.414918 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01f943c-e09c-4727-8cf7-eec58a56b363-lock\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.415058 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01f943c-e09c-4727-8cf7-eec58a56b363-cache\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.415105 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.415130 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.415164 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01f943c-e09c-4727-8cf7-eec58a56b363-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.415201 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xltnr\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-kube-api-access-xltnr\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.417022 4778 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.417050 4778 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.417099 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift podName:c01f943c-e09c-4727-8cf7-eec58a56b363 nodeName:}" failed. No retries permitted until 2026-03-12 13:31:23.917083119 +0000 UTC m=+1302.365778515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift") pod "swift-storage-0" (UID: "c01f943c-e09c-4727-8cf7-eec58a56b363") : configmap "swift-ring-files" not found Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.417101 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01f943c-e09c-4727-8cf7-eec58a56b363-lock\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.417268 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") device mount path \"/mnt/openstack/pv18\"" pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.417629 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01f943c-e09c-4727-8cf7-eec58a56b363-cache\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.422874 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01f943c-e09c-4727-8cf7-eec58a56b363-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.432193 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xltnr\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-kube-api-access-xltnr\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.437994 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.562261 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-5zdpc"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.563380 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.573320 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.573526 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.573632 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.573957 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5zdpc"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.603706 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-5zdpc"] Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.604332 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-wjwff ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-wjwff ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-5zdpc" podUID="55dae060-74ee-451e-9352-daec701140b2" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.614327 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-5knbg"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.615543 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.625858 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5knbg"] Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720656 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-combined-ca-bundle\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720702 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-dispersionconf\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720723 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-dispersionconf\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720748 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-swiftconf\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720871 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlzct\" (UniqueName: \"kubernetes.io/projected/2edc2c90-f91e-402d-809c-514e9d8a5e04-kube-api-access-jlzct\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720910 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55dae060-74ee-451e-9352-daec701140b2-etc-swift\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.720973 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-scripts\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721083 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-scripts\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721188 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2edc2c90-f91e-402d-809c-514e9d8a5e04-etc-swift\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721329 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-ring-data-devices\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721366 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-ring-data-devices\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721450 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-swiftconf\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721471 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-combined-ca-bundle\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.721503 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjwff\" (UniqueName: \"kubernetes.io/projected/55dae060-74ee-451e-9352-daec701140b2-kube-api-access-wjwff\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.823041 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-scripts\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.823104 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2edc2c90-f91e-402d-809c-514e9d8a5e04-etc-swift\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.823143 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-ring-data-devices\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.823163 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-ring-data-devices\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.823197 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-swiftconf\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.823850 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2edc2c90-f91e-402d-809c-514e9d8a5e04-etc-swift\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824040 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-scripts\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824053 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-ring-data-devices\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824051 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-ring-data-devices\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824096 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-combined-ca-bundle\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824167 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjwff\" (UniqueName: \"kubernetes.io/projected/55dae060-74ee-451e-9352-daec701140b2-kube-api-access-wjwff\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824637 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-combined-ca-bundle\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824712 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-dispersionconf\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824752 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-dispersionconf\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824816 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-swiftconf\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824874 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55dae060-74ee-451e-9352-daec701140b2-etc-swift\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824900 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlzct\" (UniqueName: \"kubernetes.io/projected/2edc2c90-f91e-402d-809c-514e9d8a5e04-kube-api-access-jlzct\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.824942 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-scripts\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.825164 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55dae060-74ee-451e-9352-daec701140b2-etc-swift\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.825628 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-scripts\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.827758 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-swiftconf\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.828814 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-dispersionconf\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.828833 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-dispersionconf\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.847940 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-combined-ca-bundle\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.848287 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-swiftconf\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.848356 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-combined-ca-bundle\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.851747 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlzct\" (UniqueName: \"kubernetes.io/projected/2edc2c90-f91e-402d-809c-514e9d8a5e04-kube-api-access-jlzct\") pod \"swift-ring-rebalance-5knbg\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.859356 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjwff\" (UniqueName: \"kubernetes.io/projected/55dae060-74ee-451e-9352-daec701140b2-kube-api-access-wjwff\") pod \"swift-ring-rebalance-5zdpc\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.926736 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.927021 4778 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.927035 4778 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 13:31:23 crc kubenswrapper[4778]: E0312 13:31:23.927080 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift podName:c01f943c-e09c-4727-8cf7-eec58a56b363 nodeName:}" failed. No retries permitted until 2026-03-12 13:31:24.9270658 +0000 UTC m=+1303.375761196 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift") pod "swift-storage-0" (UID: "c01f943c-e09c-4727-8cf7-eec58a56b363") : configmap "swift-ring-files" not found Mar 12 13:31:23 crc kubenswrapper[4778]: I0312 13:31:23.936570 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:24 crc kubenswrapper[4778]: I0312 13:31:24.087132 4778 generic.go:334] "Generic (PLEG): container finished" podID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerID="94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96" exitCode=0 Mar 12 13:31:24 crc kubenswrapper[4778]: I0312 13:31:24.088335 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8rkss" event={"ID":"5dd405d8-c82b-49d0-a871-1c7c847638df","Type":"ContainerDied","Data":"94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96"} Mar 12 13:31:24 crc kubenswrapper[4778]: I0312 13:31:24.088672 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:24 crc kubenswrapper[4778]: I0312 13:31:24.112954 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233598 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-swiftconf\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233661 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-scripts\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233703 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-combined-ca-bundle\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233725 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55dae060-74ee-451e-9352-daec701140b2-etc-swift\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233752 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-ring-data-devices\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233779 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjwff\" (UniqueName: \"kubernetes.io/projected/55dae060-74ee-451e-9352-daec701140b2-kube-api-access-wjwff\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.233806 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-dispersionconf\") pod \"55dae060-74ee-451e-9352-daec701140b2\" (UID: \"55dae060-74ee-451e-9352-daec701140b2\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.234268 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-scripts" (OuterVolumeSpecName: "scripts") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.236851 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.252324 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55dae060-74ee-451e-9352-daec701140b2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.253918 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.254631 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.259692 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.268823 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55dae060-74ee-451e-9352-daec701140b2-kube-api-access-wjwff" (OuterVolumeSpecName: "kube-api-access-wjwff") pod "55dae060-74ee-451e-9352-daec701140b2" (UID: "55dae060-74ee-451e-9352-daec701140b2"). InnerVolumeSpecName "kube-api-access-wjwff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337799 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337839 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337854 4778 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55dae060-74ee-451e-9352-daec701140b2-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337865 4778 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55dae060-74ee-451e-9352-daec701140b2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337877 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjwff\" (UniqueName: \"kubernetes.io/projected/55dae060-74ee-451e-9352-daec701140b2-kube-api-access-wjwff\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337889 4778 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.337899 4778 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55dae060-74ee-451e-9352-daec701140b2-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:24.947367 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:24.947616 4778 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:24.947892 4778 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:24.947973 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift podName:c01f943c-e09c-4727-8cf7-eec58a56b363 nodeName:}" failed. No retries permitted until 2026-03-12 13:31:26.947949716 +0000 UTC m=+1305.396645122 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift") pod "swift-storage-0" (UID: "c01f943c-e09c-4727-8cf7-eec58a56b363") : configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:25.096726 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5zdpc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:25.150232 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-5zdpc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:25.156777 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-5zdpc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.243808 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-vsbqv"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.244963 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.247225 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.263021 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55dae060-74ee-451e-9352-daec701140b2" path="/var/lib/kubelet/pods/55dae060-74ee-451e-9352-daec701140b2/volumes" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.263443 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-vsbqv"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.386960 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkg2s\" (UniqueName: \"kubernetes.io/projected/7dba49cb-b897-4877-83f4-72e0d731a1b1-kube-api-access-bkg2s\") pod \"root-account-create-update-vsbqv\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.387068 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dba49cb-b897-4877-83f4-72e0d731a1b1-operator-scripts\") pod \"root-account-create-update-vsbqv\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.489410 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkg2s\" (UniqueName: \"kubernetes.io/projected/7dba49cb-b897-4877-83f4-72e0d731a1b1-kube-api-access-bkg2s\") pod \"root-account-create-update-vsbqv\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.489476 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dba49cb-b897-4877-83f4-72e0d731a1b1-operator-scripts\") pod \"root-account-create-update-vsbqv\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.490416 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dba49cb-b897-4877-83f4-72e0d731a1b1-operator-scripts\") pod \"root-account-create-update-vsbqv\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.510736 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkg2s\" (UniqueName: \"kubernetes.io/projected/7dba49cb-b897-4877-83f4-72e0d731a1b1-kube-api-access-bkg2s\") pod \"root-account-create-update-vsbqv\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.599436 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:26.998963 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:26.999176 4778 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:26.999216 4778 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:26.999281 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift podName:c01f943c-e09c-4727-8cf7-eec58a56b363 nodeName:}" failed. No retries permitted until 2026-03-12 13:31:30.999258953 +0000 UTC m=+1309.447954349 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift") pod "swift-storage-0" (UID: "c01f943c-e09c-4727-8cf7-eec58a56b363") : configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:29.141219 4778 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Mar 12 13:31:31 crc kubenswrapper[4778]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/124fc095-41fd-4e2d-86a1-0aada5c7447f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 12 13:31:31 crc kubenswrapper[4778]: > podSandboxID="a6bf49931f5dff6309127ba2cd8c8cdab17c7cf48d3d19393ec4e377e3b30b19" Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:29.141910 4778 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 12 13:31:31 crc kubenswrapper[4778]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n599h5cbh7ch5d4h66fh676hdbh546h95h88h5ffh55ch7fhch57ch687hddhc7h5fdh57dh674h56fh64ch98h9bh557h55dh646h54ch54fh5c4h597q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5589g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-86db49b7ff-jsqnb_openstack(124fc095-41fd-4e2d-86a1-0aada5c7447f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/124fc095-41fd-4e2d-86a1-0aada5c7447f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 12 13:31:31 crc kubenswrapper[4778]: > logger="UnhandledError" Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:29.142969 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/124fc095-41fd-4e2d-86a1-0aada5c7447f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.273652 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gccjh" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.280235 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.440838 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmjgm\" (UniqueName: \"kubernetes.io/projected/199c7ab7-ef93-4b96-a76c-2476f21795ae-kube-api-access-jmjgm\") pod \"199c7ab7-ef93-4b96-a76c-2476f21795ae\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.440946 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-config\") pod \"199c7ab7-ef93-4b96-a76c-2476f21795ae\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.440978 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc051b32-4b28-4011-9a00-49caa730f074-operator-scripts\") pod \"fc051b32-4b28-4011-9a00-49caa730f074\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.441076 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-ovsdbserver-nb\") pod \"199c7ab7-ef93-4b96-a76c-2476f21795ae\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.441137 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-dns-svc\") pod \"199c7ab7-ef93-4b96-a76c-2476f21795ae\" (UID: \"199c7ab7-ef93-4b96-a76c-2476f21795ae\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.441160 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfr4k\" (UniqueName: \"kubernetes.io/projected/fc051b32-4b28-4011-9a00-49caa730f074-kube-api-access-wfr4k\") pod \"fc051b32-4b28-4011-9a00-49caa730f074\" (UID: \"fc051b32-4b28-4011-9a00-49caa730f074\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.441921 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc051b32-4b28-4011-9a00-49caa730f074-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc051b32-4b28-4011-9a00-49caa730f074" (UID: "fc051b32-4b28-4011-9a00-49caa730f074"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.454469 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc051b32-4b28-4011-9a00-49caa730f074-kube-api-access-wfr4k" (OuterVolumeSpecName: "kube-api-access-wfr4k") pod "fc051b32-4b28-4011-9a00-49caa730f074" (UID: "fc051b32-4b28-4011-9a00-49caa730f074"). InnerVolumeSpecName "kube-api-access-wfr4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.454520 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/199c7ab7-ef93-4b96-a76c-2476f21795ae-kube-api-access-jmjgm" (OuterVolumeSpecName: "kube-api-access-jmjgm") pod "199c7ab7-ef93-4b96-a76c-2476f21795ae" (UID: "199c7ab7-ef93-4b96-a76c-2476f21795ae"). InnerVolumeSpecName "kube-api-access-jmjgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.461402 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "199c7ab7-ef93-4b96-a76c-2476f21795ae" (UID: "199c7ab7-ef93-4b96-a76c-2476f21795ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.464523 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "199c7ab7-ef93-4b96-a76c-2476f21795ae" (UID: "199c7ab7-ef93-4b96-a76c-2476f21795ae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.465018 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-config" (OuterVolumeSpecName: "config") pod "199c7ab7-ef93-4b96-a76c-2476f21795ae" (UID: "199c7ab7-ef93-4b96-a76c-2476f21795ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.544300 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmjgm\" (UniqueName: \"kubernetes.io/projected/199c7ab7-ef93-4b96-a76c-2476f21795ae-kube-api-access-jmjgm\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.544335 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.544349 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc051b32-4b28-4011-9a00-49caa730f074-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.544359 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.544370 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199c7ab7-ef93-4b96-a76c-2476f21795ae-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.544380 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfr4k\" (UniqueName: \"kubernetes.io/projected/fc051b32-4b28-4011-9a00-49caa730f074-kube-api-access-wfr4k\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.602091 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-hpkvd"] Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:29.602715 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc051b32-4b28-4011-9a00-49caa730f074" containerName="mariadb-database-create" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.602740 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc051b32-4b28-4011-9a00-49caa730f074" containerName="mariadb-database-create" Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:29.602778 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199c7ab7-ef93-4b96-a76c-2476f21795ae" containerName="init" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.602790 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="199c7ab7-ef93-4b96-a76c-2476f21795ae" containerName="init" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.603043 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="199c7ab7-ef93-4b96-a76c-2476f21795ae" containerName="init" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.603065 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc051b32-4b28-4011-9a00-49caa730f074" containerName="mariadb-database-create" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.603939 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.615383 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hpkvd"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.701279 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-3148-account-create-update-zkztc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.703239 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.705737 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.713398 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3148-account-create-update-zkztc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.748769 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9cwn\" (UniqueName: \"kubernetes.io/projected/18cd7d9a-1f17-4797-a94f-4692b1180508-kube-api-access-x9cwn\") pod \"glance-db-create-hpkvd\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.749138 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18cd7d9a-1f17-4797-a94f-4692b1180508-operator-scripts\") pod \"glance-db-create-hpkvd\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.851268 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gjr5\" (UniqueName: \"kubernetes.io/projected/e76971eb-34f0-4a33-b657-508e01eed5d1-kube-api-access-4gjr5\") pod \"glance-3148-account-create-update-zkztc\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.851353 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18cd7d9a-1f17-4797-a94f-4692b1180508-operator-scripts\") pod \"glance-db-create-hpkvd\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.851821 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76971eb-34f0-4a33-b657-508e01eed5d1-operator-scripts\") pod \"glance-3148-account-create-update-zkztc\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.852089 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9cwn\" (UniqueName: \"kubernetes.io/projected/18cd7d9a-1f17-4797-a94f-4692b1180508-kube-api-access-x9cwn\") pod \"glance-db-create-hpkvd\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.852257 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18cd7d9a-1f17-4797-a94f-4692b1180508-operator-scripts\") pod \"glance-db-create-hpkvd\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.874096 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9cwn\" (UniqueName: \"kubernetes.io/projected/18cd7d9a-1f17-4797-a94f-4692b1180508-kube-api-access-x9cwn\") pod \"glance-db-create-hpkvd\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.928954 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.953354 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gjr5\" (UniqueName: \"kubernetes.io/projected/e76971eb-34f0-4a33-b657-508e01eed5d1-kube-api-access-4gjr5\") pod \"glance-3148-account-create-update-zkztc\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.953552 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76971eb-34f0-4a33-b657-508e01eed5d1-operator-scripts\") pod \"glance-3148-account-create-update-zkztc\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.954559 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76971eb-34f0-4a33-b657-508e01eed5d1-operator-scripts\") pod \"glance-3148-account-create-update-zkztc\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:29.973240 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gjr5\" (UniqueName: \"kubernetes.io/projected/e76971eb-34f0-4a33-b657-508e01eed5d1-kube-api-access-4gjr5\") pod \"glance-3148-account-create-update-zkztc\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.023826 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.142726 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8rkss" event={"ID":"5dd405d8-c82b-49d0-a871-1c7c847638df","Type":"ContainerStarted","Data":"cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb"} Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.144709 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gccjh" event={"ID":"fc051b32-4b28-4011-9a00-49caa730f074","Type":"ContainerDied","Data":"58e88409e0c9f35402bf7f7052b8bfab0894113d1308d624a95e9e78a692921f"} Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.144729 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58e88409e0c9f35402bf7f7052b8bfab0894113d1308d624a95e9e78a692921f" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.144780 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gccjh" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.154880 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.157034 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-kzfk7" event={"ID":"199c7ab7-ef93-4b96-a76c-2476f21795ae","Type":"ContainerDied","Data":"5f1e3433b2aa51609fb7a612d3a57e0e4f8b5c1392d5055466098db076137d25"} Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.157116 4778 scope.go:117] "RemoveContainer" containerID="4d8c29e71c21a3ab92dc60d0b4d6da588df8f7c85dce7dcda5be210ad42005de" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.296303 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kzfk7"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.298411 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kzfk7"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.514588 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-79rjc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.515845 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.524661 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-79rjc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.609704 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6816-account-create-update-574cj"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.611448 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.615848 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.620292 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6816-account-create-update-574cj"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.682852 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-operator-scripts\") pod \"keystone-db-create-79rjc\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.683185 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnmjq\" (UniqueName: \"kubernetes.io/projected/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-kube-api-access-fnmjq\") pod \"keystone-db-create-79rjc\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.786001 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d015b15d-96d2-4b95-9778-8f4175a840a1-operator-scripts\") pod \"keystone-6816-account-create-update-574cj\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.786217 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98q5d\" (UniqueName: \"kubernetes.io/projected/d015b15d-96d2-4b95-9778-8f4175a840a1-kube-api-access-98q5d\") pod \"keystone-6816-account-create-update-574cj\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.786310 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-operator-scripts\") pod \"keystone-db-create-79rjc\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.786349 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnmjq\" (UniqueName: \"kubernetes.io/projected/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-kube-api-access-fnmjq\") pod \"keystone-db-create-79rjc\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.787242 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-operator-scripts\") pod \"keystone-db-create-79rjc\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.814481 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnmjq\" (UniqueName: \"kubernetes.io/projected/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-kube-api-access-fnmjq\") pod \"keystone-db-create-79rjc\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.840783 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.887722 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d015b15d-96d2-4b95-9778-8f4175a840a1-operator-scripts\") pod \"keystone-6816-account-create-update-574cj\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.887821 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98q5d\" (UniqueName: \"kubernetes.io/projected/d015b15d-96d2-4b95-9778-8f4175a840a1-kube-api-access-98q5d\") pod \"keystone-6816-account-create-update-574cj\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.888806 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d015b15d-96d2-4b95-9778-8f4175a840a1-operator-scripts\") pod \"keystone-6816-account-create-update-574cj\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.903921 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98q5d\" (UniqueName: \"kubernetes.io/projected/d015b15d-96d2-4b95-9778-8f4175a840a1-kube-api-access-98q5d\") pod \"keystone-6816-account-create-update-574cj\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:30.930226 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.091383 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:31.091596 4778 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:31.091625 4778 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: E0312 13:31:31.091684 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift podName:c01f943c-e09c-4727-8cf7-eec58a56b363 nodeName:}" failed. No retries permitted until 2026-03-12 13:31:39.091664123 +0000 UTC m=+1317.540359519 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift") pod "swift-storage-0" (UID: "c01f943c-e09c-4727-8cf7-eec58a56b363") : configmap "swift-ring-files" not found Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.165425 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" event={"ID":"124fc095-41fd-4e2d-86a1-0aada5c7447f","Type":"ContainerStarted","Data":"8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d"} Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.166024 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.167399 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.195769 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" podStartSLOduration=10.631083928 podStartE2EDuration="11.195747933s" podCreationTimestamp="2026-03-12 13:31:20 +0000 UTC" firstStartedPulling="2026-03-12 13:31:21.600420575 +0000 UTC m=+1300.049115971" lastFinishedPulling="2026-03-12 13:31:22.16508458 +0000 UTC m=+1300.613779976" observedRunningTime="2026-03-12 13:31:31.186304604 +0000 UTC m=+1309.635000030" watchObservedRunningTime="2026-03-12 13:31:31.195747933 +0000 UTC m=+1309.644443339" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.215523 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-8rkss" podStartSLOduration=10.215503944 podStartE2EDuration="10.215503944s" podCreationTimestamp="2026-03-12 13:31:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:31.208737042 +0000 UTC m=+1309.657432458" watchObservedRunningTime="2026-03-12 13:31:31.215503944 +0000 UTC m=+1309.664199340" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.720527 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.908350 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b329f80-bb88-4c5c-91eb-24394cdcc492-operator-scripts\") pod \"7b329f80-bb88-4c5c-91eb-24394cdcc492\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.908465 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjpbn\" (UniqueName: \"kubernetes.io/projected/7b329f80-bb88-4c5c-91eb-24394cdcc492-kube-api-access-bjpbn\") pod \"7b329f80-bb88-4c5c-91eb-24394cdcc492\" (UID: \"7b329f80-bb88-4c5c-91eb-24394cdcc492\") " Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.909300 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b329f80-bb88-4c5c-91eb-24394cdcc492-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b329f80-bb88-4c5c-91eb-24394cdcc492" (UID: "7b329f80-bb88-4c5c-91eb-24394cdcc492"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.910956 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b329f80-bb88-4c5c-91eb-24394cdcc492-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.914976 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b329f80-bb88-4c5c-91eb-24394cdcc492-kube-api-access-bjpbn" (OuterVolumeSpecName: "kube-api-access-bjpbn") pod "7b329f80-bb88-4c5c-91eb-24394cdcc492" (UID: "7b329f80-bb88-4c5c-91eb-24394cdcc492"). InnerVolumeSpecName "kube-api-access-bjpbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.956463 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-vsbqv"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.964007 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3148-account-create-update-zkztc"] Mar 12 13:31:31 crc kubenswrapper[4778]: W0312 13:31:31.975581 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode76971eb_34f0_4a33_b657_508e01eed5d1.slice/crio-ef3bc8062b86e00dd2d40134a2cf733c4aea86e6cbc2bdc709b2200ccc77ba9a WatchSource:0}: Error finding container ef3bc8062b86e00dd2d40134a2cf733c4aea86e6cbc2bdc709b2200ccc77ba9a: Status 404 returned error can't find the container with id ef3bc8062b86e00dd2d40134a2cf733c4aea86e6cbc2bdc709b2200ccc77ba9a Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.980770 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-79rjc"] Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.987333 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5knbg"] Mar 12 13:31:31 crc kubenswrapper[4778]: W0312 13:31:31.989177 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod280f8bcd_f8e0_451d_8c9c_b733f2b62a23.slice/crio-5cdd4d026622cdc64efa7747c4f6697d32618ba64ea2916988cc0aef60712b93 WatchSource:0}: Error finding container 5cdd4d026622cdc64efa7747c4f6697d32618ba64ea2916988cc0aef60712b93: Status 404 returned error can't find the container with id 5cdd4d026622cdc64efa7747c4f6697d32618ba64ea2916988cc0aef60712b93 Mar 12 13:31:31 crc kubenswrapper[4778]: I0312 13:31:31.994749 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6816-account-create-update-574cj"] Mar 12 13:31:31 crc kubenswrapper[4778]: W0312 13:31:31.997145 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2edc2c90_f91e_402d_809c_514e9d8a5e04.slice/crio-36959ba975d8524cefb7a8390e0e734aa9fdb408ce3b4f7319c4d627f50986f6 WatchSource:0}: Error finding container 36959ba975d8524cefb7a8390e0e734aa9fdb408ce3b4f7319c4d627f50986f6: Status 404 returned error can't find the container with id 36959ba975d8524cefb7a8390e0e734aa9fdb408ce3b4f7319c4d627f50986f6 Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.002627 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hpkvd"] Mar 12 13:31:32 crc kubenswrapper[4778]: W0312 13:31:32.008759 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18cd7d9a_1f17_4797_a94f_4692b1180508.slice/crio-89192c0d3b3df7a1a0c9315e55ae046a6770c86eb0446cd32a50a504ab553bed WatchSource:0}: Error finding container 89192c0d3b3df7a1a0c9315e55ae046a6770c86eb0446cd32a50a504ab553bed: Status 404 returned error can't find the container with id 89192c0d3b3df7a1a0c9315e55ae046a6770c86eb0446cd32a50a504ab553bed Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.011827 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjpbn\" (UniqueName: \"kubernetes.io/projected/7b329f80-bb88-4c5c-91eb-24394cdcc492-kube-api-access-bjpbn\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.179253 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3148-account-create-update-zkztc" event={"ID":"e76971eb-34f0-4a33-b657-508e01eed5d1","Type":"ContainerStarted","Data":"ef3bc8062b86e00dd2d40134a2cf733c4aea86e6cbc2bdc709b2200ccc77ba9a"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.181242 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vsbqv" event={"ID":"7dba49cb-b897-4877-83f4-72e0d731a1b1","Type":"ContainerStarted","Data":"f3f7a33c33e8b6e5c107976dcfe1137727c3f5d14f498dcea6e9df482aee564a"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.181282 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vsbqv" event={"ID":"7dba49cb-b897-4877-83f4-72e0d731a1b1","Type":"ContainerStarted","Data":"bedc9fba69e3f28bbfd7d4f69a0cadd95acf08c6bd39f0e86608134517a0bb5c"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.184573 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6816-account-create-update-574cj" event={"ID":"d015b15d-96d2-4b95-9778-8f4175a840a1","Type":"ContainerStarted","Data":"f4bb8c6e00b5e03bcc01c6649d1104fc5ef38426458fa36f98588fb6167dbe07"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.184640 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6816-account-create-update-574cj" event={"ID":"d015b15d-96d2-4b95-9778-8f4175a840a1","Type":"ContainerStarted","Data":"9bb61e64fae43b699f3efd9be919423ad2bf3a2faceb2aa570b27cb8710da6c0"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.186239 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-79rjc" event={"ID":"280f8bcd-f8e0-451d-8c9c-b733f2b62a23","Type":"ContainerStarted","Data":"5cdd4d026622cdc64efa7747c4f6697d32618ba64ea2916988cc0aef60712b93"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.188792 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5knbg" event={"ID":"2edc2c90-f91e-402d-809c-514e9d8a5e04","Type":"ContainerStarted","Data":"36959ba975d8524cefb7a8390e0e734aa9fdb408ce3b4f7319c4d627f50986f6"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.190936 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hpkvd" event={"ID":"18cd7d9a-1f17-4797-a94f-4692b1180508","Type":"ContainerStarted","Data":"89192c0d3b3df7a1a0c9315e55ae046a6770c86eb0446cd32a50a504ab553bed"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.193921 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b25f9c9-784a-4a52-9bb3-02c6c4592702","Type":"ContainerStarted","Data":"f8ecf9c7642de7714bf2b00724fe551c713a4b7cdd00502f8c955ee95a91b067"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.193960 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b25f9c9-784a-4a52-9bb3-02c6c4592702","Type":"ContainerStarted","Data":"1aabf1656062aeb27eb7042fd97a5654699971b77c5202dbaf7a6937510839a0"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.193975 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.199619 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-vsbqv" podStartSLOduration=6.199600865 podStartE2EDuration="6.199600865s" podCreationTimestamp="2026-03-12 13:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:32.198306728 +0000 UTC m=+1310.647002124" watchObservedRunningTime="2026-03-12 13:31:32.199600865 +0000 UTC m=+1310.648296261" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.201226 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-886c-account-create-update-c7kqb" event={"ID":"7b329f80-bb88-4c5c-91eb-24394cdcc492","Type":"ContainerDied","Data":"4931bee01e54e487e69bca458cb118194d8730ef5cb3f2125d81402dd58d404b"} Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.201275 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4931bee01e54e487e69bca458cb118194d8730ef5cb3f2125d81402dd58d404b" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.201251 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-886c-account-create-update-c7kqb" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.231280 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.112331773 podStartE2EDuration="11.231256915s" podCreationTimestamp="2026-03-12 13:31:21 +0000 UTC" firstStartedPulling="2026-03-12 13:31:22.162365753 +0000 UTC m=+1300.611061149" lastFinishedPulling="2026-03-12 13:31:31.281290875 +0000 UTC m=+1309.729986291" observedRunningTime="2026-03-12 13:31:32.217114223 +0000 UTC m=+1310.665809639" watchObservedRunningTime="2026-03-12 13:31:32.231256915 +0000 UTC m=+1310.679952501" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.260683 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-79rjc" podStartSLOduration=2.260669311 podStartE2EDuration="2.260669311s" podCreationTimestamp="2026-03-12 13:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:32.247997641 +0000 UTC m=+1310.696693037" watchObservedRunningTime="2026-03-12 13:31:32.260669311 +0000 UTC m=+1310.709364697" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.274911 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6816-account-create-update-574cj" podStartSLOduration=2.274893216 podStartE2EDuration="2.274893216s" podCreationTimestamp="2026-03-12 13:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:32.269442621 +0000 UTC m=+1310.718138017" watchObservedRunningTime="2026-03-12 13:31:32.274893216 +0000 UTC m=+1310.723588612" Mar 12 13:31:32 crc kubenswrapper[4778]: I0312 13:31:32.298115 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="199c7ab7-ef93-4b96-a76c-2476f21795ae" path="/var/lib/kubelet/pods/199c7ab7-ef93-4b96-a76c-2476f21795ae/volumes" Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.209569 4778 generic.go:334] "Generic (PLEG): container finished" podID="7dba49cb-b897-4877-83f4-72e0d731a1b1" containerID="f3f7a33c33e8b6e5c107976dcfe1137727c3f5d14f498dcea6e9df482aee564a" exitCode=0 Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.209645 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vsbqv" event={"ID":"7dba49cb-b897-4877-83f4-72e0d731a1b1","Type":"ContainerDied","Data":"f3f7a33c33e8b6e5c107976dcfe1137727c3f5d14f498dcea6e9df482aee564a"} Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.212417 4778 generic.go:334] "Generic (PLEG): container finished" podID="d015b15d-96d2-4b95-9778-8f4175a840a1" containerID="f4bb8c6e00b5e03bcc01c6649d1104fc5ef38426458fa36f98588fb6167dbe07" exitCode=0 Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.212551 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6816-account-create-update-574cj" event={"ID":"d015b15d-96d2-4b95-9778-8f4175a840a1","Type":"ContainerDied","Data":"f4bb8c6e00b5e03bcc01c6649d1104fc5ef38426458fa36f98588fb6167dbe07"} Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.214799 4778 generic.go:334] "Generic (PLEG): container finished" podID="280f8bcd-f8e0-451d-8c9c-b733f2b62a23" containerID="8443f2894188b4c3d976d78d2d647409527ab07f04b215d8b647fc560059ba2f" exitCode=0 Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.214896 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-79rjc" event={"ID":"280f8bcd-f8e0-451d-8c9c-b733f2b62a23","Type":"ContainerDied","Data":"8443f2894188b4c3d976d78d2d647409527ab07f04b215d8b647fc560059ba2f"} Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.215974 4778 generic.go:334] "Generic (PLEG): container finished" podID="18cd7d9a-1f17-4797-a94f-4692b1180508" containerID="451301ebd2071510b670f3a924d5fcd2f28fbcc4aa60d4224906bca0e09aa5be" exitCode=0 Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.216022 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hpkvd" event={"ID":"18cd7d9a-1f17-4797-a94f-4692b1180508","Type":"ContainerDied","Data":"451301ebd2071510b670f3a924d5fcd2f28fbcc4aa60d4224906bca0e09aa5be"} Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.219948 4778 generic.go:334] "Generic (PLEG): container finished" podID="e76971eb-34f0-4a33-b657-508e01eed5d1" containerID="5106184b767437cea31a6a61b3a1991b36587ddd28250ecc1207af703f368fda" exitCode=0 Mar 12 13:31:33 crc kubenswrapper[4778]: I0312 13:31:33.220042 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3148-account-create-update-zkztc" event={"ID":"e76971eb-34f0-4a33-b657-508e01eed5d1","Type":"ContainerDied","Data":"5106184b767437cea31a6a61b3a1991b36587ddd28250ecc1207af703f368fda"} Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.159204 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.167147 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.189674 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dba49cb-b897-4877-83f4-72e0d731a1b1-operator-scripts\") pod \"7dba49cb-b897-4877-83f4-72e0d731a1b1\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.189800 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkg2s\" (UniqueName: \"kubernetes.io/projected/7dba49cb-b897-4877-83f4-72e0d731a1b1-kube-api-access-bkg2s\") pod \"7dba49cb-b897-4877-83f4-72e0d731a1b1\" (UID: \"7dba49cb-b897-4877-83f4-72e0d731a1b1\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.189835 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.194826 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.194849 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dba49cb-b897-4877-83f4-72e0d731a1b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7dba49cb-b897-4877-83f4-72e0d731a1b1" (UID: "7dba49cb-b897-4877-83f4-72e0d731a1b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.207216 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dba49cb-b897-4877-83f4-72e0d731a1b1-kube-api-access-bkg2s" (OuterVolumeSpecName: "kube-api-access-bkg2s") pod "7dba49cb-b897-4877-83f4-72e0d731a1b1" (UID: "7dba49cb-b897-4877-83f4-72e0d731a1b1"). InnerVolumeSpecName "kube-api-access-bkg2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.211162 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.261619 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3148-account-create-update-zkztc" event={"ID":"e76971eb-34f0-4a33-b657-508e01eed5d1","Type":"ContainerDied","Data":"ef3bc8062b86e00dd2d40134a2cf733c4aea86e6cbc2bdc709b2200ccc77ba9a"} Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.261663 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef3bc8062b86e00dd2d40134a2cf733c4aea86e6cbc2bdc709b2200ccc77ba9a" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.261772 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3148-account-create-update-zkztc" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.263493 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vsbqv" event={"ID":"7dba49cb-b897-4877-83f4-72e0d731a1b1","Type":"ContainerDied","Data":"bedc9fba69e3f28bbfd7d4f69a0cadd95acf08c6bd39f0e86608134517a0bb5c"} Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.263517 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bedc9fba69e3f28bbfd7d4f69a0cadd95acf08c6bd39f0e86608134517a0bb5c" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.263583 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vsbqv" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.274156 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6816-account-create-update-574cj" event={"ID":"d015b15d-96d2-4b95-9778-8f4175a840a1","Type":"ContainerDied","Data":"9bb61e64fae43b699f3efd9be919423ad2bf3a2faceb2aa570b27cb8710da6c0"} Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.274244 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bb61e64fae43b699f3efd9be919423ad2bf3a2faceb2aa570b27cb8710da6c0" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.274295 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6816-account-create-update-574cj" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.276118 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-79rjc" event={"ID":"280f8bcd-f8e0-451d-8c9c-b733f2b62a23","Type":"ContainerDied","Data":"5cdd4d026622cdc64efa7747c4f6697d32618ba64ea2916988cc0aef60712b93"} Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.276140 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-79rjc" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.276142 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cdd4d026622cdc64efa7747c4f6697d32618ba64ea2916988cc0aef60712b93" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.277144 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hpkvd" event={"ID":"18cd7d9a-1f17-4797-a94f-4692b1180508","Type":"ContainerDied","Data":"89192c0d3b3df7a1a0c9315e55ae046a6770c86eb0446cd32a50a504ab553bed"} Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.277163 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89192c0d3b3df7a1a0c9315e55ae046a6770c86eb0446cd32a50a504ab553bed" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.277167 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hpkvd" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.292918 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gjr5\" (UniqueName: \"kubernetes.io/projected/e76971eb-34f0-4a33-b657-508e01eed5d1-kube-api-access-4gjr5\") pod \"e76971eb-34f0-4a33-b657-508e01eed5d1\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.292992 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnmjq\" (UniqueName: \"kubernetes.io/projected/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-kube-api-access-fnmjq\") pod \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.293030 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18cd7d9a-1f17-4797-a94f-4692b1180508-operator-scripts\") pod \"18cd7d9a-1f17-4797-a94f-4692b1180508\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.293106 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9cwn\" (UniqueName: \"kubernetes.io/projected/18cd7d9a-1f17-4797-a94f-4692b1180508-kube-api-access-x9cwn\") pod \"18cd7d9a-1f17-4797-a94f-4692b1180508\" (UID: \"18cd7d9a-1f17-4797-a94f-4692b1180508\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.293143 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98q5d\" (UniqueName: \"kubernetes.io/projected/d015b15d-96d2-4b95-9778-8f4175a840a1-kube-api-access-98q5d\") pod \"d015b15d-96d2-4b95-9778-8f4175a840a1\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.293591 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18cd7d9a-1f17-4797-a94f-4692b1180508-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18cd7d9a-1f17-4797-a94f-4692b1180508" (UID: "18cd7d9a-1f17-4797-a94f-4692b1180508"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.296741 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-operator-scripts\") pod \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\" (UID: \"280f8bcd-f8e0-451d-8c9c-b733f2b62a23\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.296822 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76971eb-34f0-4a33-b657-508e01eed5d1-operator-scripts\") pod \"e76971eb-34f0-4a33-b657-508e01eed5d1\" (UID: \"e76971eb-34f0-4a33-b657-508e01eed5d1\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.296806 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d015b15d-96d2-4b95-9778-8f4175a840a1-kube-api-access-98q5d" (OuterVolumeSpecName: "kube-api-access-98q5d") pod "d015b15d-96d2-4b95-9778-8f4175a840a1" (UID: "d015b15d-96d2-4b95-9778-8f4175a840a1"). InnerVolumeSpecName "kube-api-access-98q5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.296860 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d015b15d-96d2-4b95-9778-8f4175a840a1-operator-scripts\") pod \"d015b15d-96d2-4b95-9778-8f4175a840a1\" (UID: \"d015b15d-96d2-4b95-9778-8f4175a840a1\") " Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.297266 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "280f8bcd-f8e0-451d-8c9c-b733f2b62a23" (UID: "280f8bcd-f8e0-451d-8c9c-b733f2b62a23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.297279 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e76971eb-34f0-4a33-b657-508e01eed5d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e76971eb-34f0-4a33-b657-508e01eed5d1" (UID: "e76971eb-34f0-4a33-b657-508e01eed5d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.297347 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d015b15d-96d2-4b95-9778-8f4175a840a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d015b15d-96d2-4b95-9778-8f4175a840a1" (UID: "d015b15d-96d2-4b95-9778-8f4175a840a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.297505 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76971eb-34f0-4a33-b657-508e01eed5d1-kube-api-access-4gjr5" (OuterVolumeSpecName: "kube-api-access-4gjr5") pod "e76971eb-34f0-4a33-b657-508e01eed5d1" (UID: "e76971eb-34f0-4a33-b657-508e01eed5d1"). InnerVolumeSpecName "kube-api-access-4gjr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298521 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298577 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dba49cb-b897-4877-83f4-72e0d731a1b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298596 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76971eb-34f0-4a33-b657-508e01eed5d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298610 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d015b15d-96d2-4b95-9778-8f4175a840a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298623 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkg2s\" (UniqueName: \"kubernetes.io/projected/7dba49cb-b897-4877-83f4-72e0d731a1b1-kube-api-access-bkg2s\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298679 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gjr5\" (UniqueName: \"kubernetes.io/projected/e76971eb-34f0-4a33-b657-508e01eed5d1-kube-api-access-4gjr5\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298704 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18cd7d9a-1f17-4797-a94f-4692b1180508-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298751 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98q5d\" (UniqueName: \"kubernetes.io/projected/d015b15d-96d2-4b95-9778-8f4175a840a1-kube-api-access-98q5d\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.298874 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-kube-api-access-fnmjq" (OuterVolumeSpecName: "kube-api-access-fnmjq") pod "280f8bcd-f8e0-451d-8c9c-b733f2b62a23" (UID: "280f8bcd-f8e0-451d-8c9c-b733f2b62a23"). InnerVolumeSpecName "kube-api-access-fnmjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.300393 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18cd7d9a-1f17-4797-a94f-4692b1180508-kube-api-access-x9cwn" (OuterVolumeSpecName: "kube-api-access-x9cwn") pod "18cd7d9a-1f17-4797-a94f-4692b1180508" (UID: "18cd7d9a-1f17-4797-a94f-4692b1180508"). InnerVolumeSpecName "kube-api-access-x9cwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.399560 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnmjq\" (UniqueName: \"kubernetes.io/projected/280f8bcd-f8e0-451d-8c9c-b733f2b62a23-kube-api-access-fnmjq\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:35 crc kubenswrapper[4778]: I0312 13:31:35.399629 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9cwn\" (UniqueName: \"kubernetes.io/projected/18cd7d9a-1f17-4797-a94f-4692b1180508-kube-api-access-x9cwn\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:36 crc kubenswrapper[4778]: I0312 13:31:36.286382 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5knbg" event={"ID":"2edc2c90-f91e-402d-809c-514e9d8a5e04","Type":"ContainerStarted","Data":"112c5296361c82469b890fc71a2c6b309a06a72b7d67b5062a1ead56745507c2"} Mar 12 13:31:36 crc kubenswrapper[4778]: I0312 13:31:36.305340 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-5knbg" podStartSLOduration=10.293351913 podStartE2EDuration="13.305315174s" podCreationTimestamp="2026-03-12 13:31:23 +0000 UTC" firstStartedPulling="2026-03-12 13:31:31.999566767 +0000 UTC m=+1310.448262173" lastFinishedPulling="2026-03-12 13:31:35.011530028 +0000 UTC m=+1313.460225434" observedRunningTime="2026-03-12 13:31:36.301244078 +0000 UTC m=+1314.749939484" watchObservedRunningTime="2026-03-12 13:31:36.305315174 +0000 UTC m=+1314.754010570" Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.364409 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.422501 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jsqnb"] Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.422770 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerName="dnsmasq-dns" containerID="cri-o://8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d" gracePeriod=10 Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.424452 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.746073 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-vsbqv"] Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.755500 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-vsbqv"] Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.849357 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.939930 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-config\") pod \"124fc095-41fd-4e2d-86a1-0aada5c7447f\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.940245 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-nb\") pod \"124fc095-41fd-4e2d-86a1-0aada5c7447f\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.940466 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-dns-svc\") pod \"124fc095-41fd-4e2d-86a1-0aada5c7447f\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.940884 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5589g\" (UniqueName: \"kubernetes.io/projected/124fc095-41fd-4e2d-86a1-0aada5c7447f-kube-api-access-5589g\") pod \"124fc095-41fd-4e2d-86a1-0aada5c7447f\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.944605 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-sb\") pod \"124fc095-41fd-4e2d-86a1-0aada5c7447f\" (UID: \"124fc095-41fd-4e2d-86a1-0aada5c7447f\") " Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.973511 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/124fc095-41fd-4e2d-86a1-0aada5c7447f-kube-api-access-5589g" (OuterVolumeSpecName: "kube-api-access-5589g") pod "124fc095-41fd-4e2d-86a1-0aada5c7447f" (UID: "124fc095-41fd-4e2d-86a1-0aada5c7447f"). InnerVolumeSpecName "kube-api-access-5589g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.990719 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "124fc095-41fd-4e2d-86a1-0aada5c7447f" (UID: "124fc095-41fd-4e2d-86a1-0aada5c7447f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:37 crc kubenswrapper[4778]: I0312 13:31:37.993344 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "124fc095-41fd-4e2d-86a1-0aada5c7447f" (UID: "124fc095-41fd-4e2d-86a1-0aada5c7447f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.007643 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-config" (OuterVolumeSpecName: "config") pod "124fc095-41fd-4e2d-86a1-0aada5c7447f" (UID: "124fc095-41fd-4e2d-86a1-0aada5c7447f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.008341 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "124fc095-41fd-4e2d-86a1-0aada5c7447f" (UID: "124fc095-41fd-4e2d-86a1-0aada5c7447f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.052881 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.052927 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5589g\" (UniqueName: \"kubernetes.io/projected/124fc095-41fd-4e2d-86a1-0aada5c7447f-kube-api-access-5589g\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.052944 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.052957 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.052968 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/124fc095-41fd-4e2d-86a1-0aada5c7447f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.261981 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dba49cb-b897-4877-83f4-72e0d731a1b1" path="/var/lib/kubelet/pods/7dba49cb-b897-4877-83f4-72e0d731a1b1/volumes" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.301656 4778 generic.go:334] "Generic (PLEG): container finished" podID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerID="8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d" exitCode=0 Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.301705 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" event={"ID":"124fc095-41fd-4e2d-86a1-0aada5c7447f","Type":"ContainerDied","Data":"8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d"} Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.301720 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.301730 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-jsqnb" event={"ID":"124fc095-41fd-4e2d-86a1-0aada5c7447f","Type":"ContainerDied","Data":"a6bf49931f5dff6309127ba2cd8c8cdab17c7cf48d3d19393ec4e377e3b30b19"} Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.301746 4778 scope.go:117] "RemoveContainer" containerID="8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.321561 4778 scope.go:117] "RemoveContainer" containerID="2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.328650 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jsqnb"] Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.334022 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-jsqnb"] Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.338959 4778 scope.go:117] "RemoveContainer" containerID="8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d" Mar 12 13:31:38 crc kubenswrapper[4778]: E0312 13:31:38.340239 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d\": container with ID starting with 8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d not found: ID does not exist" containerID="8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.340290 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d"} err="failed to get container status \"8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d\": rpc error: code = NotFound desc = could not find container \"8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d\": container with ID starting with 8df5a4af2891ad2a0c3dc58f39282cbd926977260f1e62c77d95d627bea99c7d not found: ID does not exist" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.340316 4778 scope.go:117] "RemoveContainer" containerID="2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9" Mar 12 13:31:38 crc kubenswrapper[4778]: E0312 13:31:38.340575 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9\": container with ID starting with 2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9 not found: ID does not exist" containerID="2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9" Mar 12 13:31:38 crc kubenswrapper[4778]: I0312 13:31:38.340604 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9"} err="failed to get container status \"2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9\": rpc error: code = NotFound desc = could not find container \"2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9\": container with ID starting with 2c4c2a78bfb0c622d1943285b717a26c265f52cecb8566f14114c1ff4b03e4c9 not found: ID does not exist" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.174267 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.174544 4778 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.174574 4778 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.174646 4778 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift podName:c01f943c-e09c-4727-8cf7-eec58a56b363 nodeName:}" failed. No retries permitted until 2026-03-12 13:31:55.174623028 +0000 UTC m=+1333.623318434 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift") pod "swift-storage-0" (UID: "c01f943c-e09c-4727-8cf7-eec58a56b363") : configmap "swift-ring-files" not found Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868283 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-xg6z4"] Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868771 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d015b15d-96d2-4b95-9778-8f4175a840a1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868796 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d015b15d-96d2-4b95-9778-8f4175a840a1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868825 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cd7d9a-1f17-4797-a94f-4692b1180508" containerName="mariadb-database-create" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868833 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cd7d9a-1f17-4797-a94f-4692b1180508" containerName="mariadb-database-create" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868851 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerName="init" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868860 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerName="init" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868873 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280f8bcd-f8e0-451d-8c9c-b733f2b62a23" containerName="mariadb-database-create" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868884 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="280f8bcd-f8e0-451d-8c9c-b733f2b62a23" containerName="mariadb-database-create" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868897 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b329f80-bb88-4c5c-91eb-24394cdcc492" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868905 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b329f80-bb88-4c5c-91eb-24394cdcc492" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868921 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerName="dnsmasq-dns" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868928 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerName="dnsmasq-dns" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868951 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76971eb-34f0-4a33-b657-508e01eed5d1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868958 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76971eb-34f0-4a33-b657-508e01eed5d1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: E0312 13:31:39.868971 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dba49cb-b897-4877-83f4-72e0d731a1b1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.868979 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dba49cb-b897-4877-83f4-72e0d731a1b1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869177 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76971eb-34f0-4a33-b657-508e01eed5d1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869233 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cd7d9a-1f17-4797-a94f-4692b1180508" containerName="mariadb-database-create" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869242 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d015b15d-96d2-4b95-9778-8f4175a840a1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869255 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" containerName="dnsmasq-dns" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869267 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b329f80-bb88-4c5c-91eb-24394cdcc492" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869278 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dba49cb-b897-4877-83f4-72e0d731a1b1" containerName="mariadb-account-create-update" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869291 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="280f8bcd-f8e0-451d-8c9c-b733f2b62a23" containerName="mariadb-database-create" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.869859 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.873091 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.873114 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l7l5j" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.901011 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-xg6z4"] Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.988222 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-config-data\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.988299 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-db-sync-config-data\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.988362 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-combined-ca-bundle\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:39 crc kubenswrapper[4778]: I0312 13:31:39.988420 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crgvn\" (UniqueName: \"kubernetes.io/projected/befeb973-a1de-48f9-8de0-5559f75472dc-kube-api-access-crgvn\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.090119 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-config-data\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.090207 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-db-sync-config-data\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.090269 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-combined-ca-bundle\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.090300 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crgvn\" (UniqueName: \"kubernetes.io/projected/befeb973-a1de-48f9-8de0-5559f75472dc-kube-api-access-crgvn\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.095494 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-db-sync-config-data\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.095667 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-combined-ca-bundle\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.095779 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-config-data\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.108431 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crgvn\" (UniqueName: \"kubernetes.io/projected/befeb973-a1de-48f9-8de0-5559f75472dc-kube-api-access-crgvn\") pod \"glance-db-sync-xg6z4\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.190058 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xg6z4" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.270506 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="124fc095-41fd-4e2d-86a1-0aada5c7447f" path="/var/lib/kubelet/pods/124fc095-41fd-4e2d-86a1-0aada5c7447f/volumes" Mar 12 13:31:40 crc kubenswrapper[4778]: I0312 13:31:40.715366 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-xg6z4"] Mar 12 13:31:40 crc kubenswrapper[4778]: W0312 13:31:40.736645 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbefeb973_a1de_48f9_8de0_5559f75472dc.slice/crio-f4635ea2bc5d2d0cce58645ef33f0143795167ef564ca2829fbc3740cec61b52 WatchSource:0}: Error finding container f4635ea2bc5d2d0cce58645ef33f0143795167ef564ca2829fbc3740cec61b52: Status 404 returned error can't find the container with id f4635ea2bc5d2d0cce58645ef33f0143795167ef564ca2829fbc3740cec61b52 Mar 12 13:31:41 crc kubenswrapper[4778]: I0312 13:31:41.328831 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xg6z4" event={"ID":"befeb973-a1de-48f9-8de0-5559f75472dc","Type":"ContainerStarted","Data":"f4635ea2bc5d2d0cce58645ef33f0143795167ef564ca2829fbc3740cec61b52"} Mar 12 13:31:41 crc kubenswrapper[4778]: I0312 13:31:41.495174 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.344708 4778 generic.go:334] "Generic (PLEG): container finished" podID="2edc2c90-f91e-402d-809c-514e9d8a5e04" containerID="112c5296361c82469b890fc71a2c6b309a06a72b7d67b5062a1ead56745507c2" exitCode=0 Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.344763 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5knbg" event={"ID":"2edc2c90-f91e-402d-809c-514e9d8a5e04","Type":"ContainerDied","Data":"112c5296361c82469b890fc71a2c6b309a06a72b7d67b5062a1ead56745507c2"} Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.773307 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-7kt6z"] Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.774736 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.776620 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.779166 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-7kt6z"] Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.843939 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf8gm\" (UniqueName: \"kubernetes.io/projected/dd5a0cd9-113c-4313-8d66-90487bd90cd3-kube-api-access-wf8gm\") pod \"root-account-create-update-7kt6z\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.844064 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd5a0cd9-113c-4313-8d66-90487bd90cd3-operator-scripts\") pod \"root-account-create-update-7kt6z\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.946104 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd5a0cd9-113c-4313-8d66-90487bd90cd3-operator-scripts\") pod \"root-account-create-update-7kt6z\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.946244 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf8gm\" (UniqueName: \"kubernetes.io/projected/dd5a0cd9-113c-4313-8d66-90487bd90cd3-kube-api-access-wf8gm\") pod \"root-account-create-update-7kt6z\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.946936 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd5a0cd9-113c-4313-8d66-90487bd90cd3-operator-scripts\") pod \"root-account-create-update-7kt6z\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:42 crc kubenswrapper[4778]: I0312 13:31:42.964076 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf8gm\" (UniqueName: \"kubernetes.io/projected/dd5a0cd9-113c-4313-8d66-90487bd90cd3-kube-api-access-wf8gm\") pod \"root-account-create-update-7kt6z\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.091962 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.539619 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-7kt6z"] Mar 12 13:31:43 crc kubenswrapper[4778]: W0312 13:31:43.555205 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd5a0cd9_113c_4313_8d66_90487bd90cd3.slice/crio-052938a160ec540bb5a700bb99d2b7290aaa6a277baedbccdd9c5a8f16111da3 WatchSource:0}: Error finding container 052938a160ec540bb5a700bb99d2b7290aaa6a277baedbccdd9c5a8f16111da3: Status 404 returned error can't find the container with id 052938a160ec540bb5a700bb99d2b7290aaa6a277baedbccdd9c5a8f16111da3 Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.668410 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.760960 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-ring-data-devices\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761537 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-dispersionconf\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761625 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-scripts\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761699 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlzct\" (UniqueName: \"kubernetes.io/projected/2edc2c90-f91e-402d-809c-514e9d8a5e04-kube-api-access-jlzct\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761718 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-combined-ca-bundle\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761773 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761784 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2edc2c90-f91e-402d-809c-514e9d8a5e04-etc-swift\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.761868 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-swiftconf\") pod \"2edc2c90-f91e-402d-809c-514e9d8a5e04\" (UID: \"2edc2c90-f91e-402d-809c-514e9d8a5e04\") " Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.762475 4778 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.763172 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2edc2c90-f91e-402d-809c-514e9d8a5e04-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.770065 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2edc2c90-f91e-402d-809c-514e9d8a5e04-kube-api-access-jlzct" (OuterVolumeSpecName: "kube-api-access-jlzct") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "kube-api-access-jlzct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.777619 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.792688 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-scripts" (OuterVolumeSpecName: "scripts") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.800018 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.800779 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2edc2c90-f91e-402d-809c-514e9d8a5e04" (UID: "2edc2c90-f91e-402d-809c-514e9d8a5e04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.864608 4778 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.864655 4778 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.864672 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edc2c90-f91e-402d-809c-514e9d8a5e04-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.864755 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlzct\" (UniqueName: \"kubernetes.io/projected/2edc2c90-f91e-402d-809c-514e9d8a5e04-kube-api-access-jlzct\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.864771 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edc2c90-f91e-402d-809c-514e9d8a5e04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:43 crc kubenswrapper[4778]: I0312 13:31:43.864785 4778 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2edc2c90-f91e-402d-809c-514e9d8a5e04-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.377362 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5knbg" event={"ID":"2edc2c90-f91e-402d-809c-514e9d8a5e04","Type":"ContainerDied","Data":"36959ba975d8524cefb7a8390e0e734aa9fdb408ce3b4f7319c4d627f50986f6"} Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.377434 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36959ba975d8524cefb7a8390e0e734aa9fdb408ce3b4f7319c4d627f50986f6" Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.377531 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5knbg" Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.394137 4778 generic.go:334] "Generic (PLEG): container finished" podID="dd5a0cd9-113c-4313-8d66-90487bd90cd3" containerID="af7a0409b1470d33d558b70c98a397f0b5c99782ac9578ab1f379f9cb685947f" exitCode=0 Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.395153 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7kt6z" event={"ID":"dd5a0cd9-113c-4313-8d66-90487bd90cd3","Type":"ContainerDied","Data":"af7a0409b1470d33d558b70c98a397f0b5c99782ac9578ab1f379f9cb685947f"} Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.399814 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7kt6z" event={"ID":"dd5a0cd9-113c-4313-8d66-90487bd90cd3","Type":"ContainerStarted","Data":"052938a160ec540bb5a700bb99d2b7290aaa6a277baedbccdd9c5a8f16111da3"} Mar 12 13:31:44 crc kubenswrapper[4778]: I0312 13:31:44.797868 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4wct6" podUID="3b8efd1e-884d-4963-b69f-04ede0a92267" containerName="ovn-controller" probeResult="failure" output=< Mar 12 13:31:44 crc kubenswrapper[4778]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 12 13:31:44 crc kubenswrapper[4778]: > Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.749836 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.796872 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf8gm\" (UniqueName: \"kubernetes.io/projected/dd5a0cd9-113c-4313-8d66-90487bd90cd3-kube-api-access-wf8gm\") pod \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.796947 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd5a0cd9-113c-4313-8d66-90487bd90cd3-operator-scripts\") pod \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\" (UID: \"dd5a0cd9-113c-4313-8d66-90487bd90cd3\") " Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.797792 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd5a0cd9-113c-4313-8d66-90487bd90cd3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd5a0cd9-113c-4313-8d66-90487bd90cd3" (UID: "dd5a0cd9-113c-4313-8d66-90487bd90cd3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.802010 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd5a0cd9-113c-4313-8d66-90487bd90cd3-kube-api-access-wf8gm" (OuterVolumeSpecName: "kube-api-access-wf8gm") pod "dd5a0cd9-113c-4313-8d66-90487bd90cd3" (UID: "dd5a0cd9-113c-4313-8d66-90487bd90cd3"). InnerVolumeSpecName "kube-api-access-wf8gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.899303 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf8gm\" (UniqueName: \"kubernetes.io/projected/dd5a0cd9-113c-4313-8d66-90487bd90cd3-kube-api-access-wf8gm\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:45 crc kubenswrapper[4778]: I0312 13:31:45.899337 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd5a0cd9-113c-4313-8d66-90487bd90cd3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:31:46 crc kubenswrapper[4778]: I0312 13:31:46.417332 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7kt6z" event={"ID":"dd5a0cd9-113c-4313-8d66-90487bd90cd3","Type":"ContainerDied","Data":"052938a160ec540bb5a700bb99d2b7290aaa6a277baedbccdd9c5a8f16111da3"} Mar 12 13:31:46 crc kubenswrapper[4778]: I0312 13:31:46.417738 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="052938a160ec540bb5a700bb99d2b7290aaa6a277baedbccdd9c5a8f16111da3" Mar 12 13:31:46 crc kubenswrapper[4778]: I0312 13:31:46.417405 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7kt6z" Mar 12 13:31:49 crc kubenswrapper[4778]: I0312 13:31:49.444764 4778 generic.go:334] "Generic (PLEG): container finished" podID="1e89dfcc-2ac3-444c-91e8-56991eae096b" containerID="491cf83ea2b0803c619e4110e5a18dd9c9b6e2cc2bfd596357f59a6a18312dee" exitCode=0 Mar 12 13:31:49 crc kubenswrapper[4778]: I0312 13:31:49.444863 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e89dfcc-2ac3-444c-91e8-56991eae096b","Type":"ContainerDied","Data":"491cf83ea2b0803c619e4110e5a18dd9c9b6e2cc2bfd596357f59a6a18312dee"} Mar 12 13:31:49 crc kubenswrapper[4778]: I0312 13:31:49.792250 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4wct6" podUID="3b8efd1e-884d-4963-b69f-04ede0a92267" containerName="ovn-controller" probeResult="failure" output=< Mar 12 13:31:49 crc kubenswrapper[4778]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 12 13:31:49 crc kubenswrapper[4778]: > Mar 12 13:31:49 crc kubenswrapper[4778]: I0312 13:31:49.813512 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:31:49 crc kubenswrapper[4778]: I0312 13:31:49.815170 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-p67vh" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.069720 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4wct6-config-6fwv4"] Mar 12 13:31:50 crc kubenswrapper[4778]: E0312 13:31:50.070441 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edc2c90-f91e-402d-809c-514e9d8a5e04" containerName="swift-ring-rebalance" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.070471 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edc2c90-f91e-402d-809c-514e9d8a5e04" containerName="swift-ring-rebalance" Mar 12 13:31:50 crc kubenswrapper[4778]: E0312 13:31:50.070485 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5a0cd9-113c-4313-8d66-90487bd90cd3" containerName="mariadb-account-create-update" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.070495 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5a0cd9-113c-4313-8d66-90487bd90cd3" containerName="mariadb-account-create-update" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.070774 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edc2c90-f91e-402d-809c-514e9d8a5e04" containerName="swift-ring-rebalance" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.070821 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd5a0cd9-113c-4313-8d66-90487bd90cd3" containerName="mariadb-account-create-update" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.071962 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.074874 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.078394 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4wct6-config-6fwv4"] Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.186682 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-additional-scripts\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.186737 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-log-ovn\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.186764 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run-ovn\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.186861 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.186899 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t2nc\" (UniqueName: \"kubernetes.io/projected/20836760-c025-4e65-bf24-34fc17f3c649-kube-api-access-4t2nc\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.187030 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-scripts\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289036 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-scripts\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289109 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-additional-scripts\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289138 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-log-ovn\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289153 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run-ovn\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289196 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289249 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t2nc\" (UniqueName: \"kubernetes.io/projected/20836760-c025-4e65-bf24-34fc17f3c649-kube-api-access-4t2nc\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289673 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run-ovn\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.289691 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-log-ovn\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.290396 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-additional-scripts\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.290688 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.293927 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-scripts\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.335017 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t2nc\" (UniqueName: \"kubernetes.io/projected/20836760-c025-4e65-bf24-34fc17f3c649-kube-api-access-4t2nc\") pod \"ovn-controller-4wct6-config-6fwv4\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:50 crc kubenswrapper[4778]: I0312 13:31:50.391555 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:31:52 crc kubenswrapper[4778]: I0312 13:31:52.483169 4778 generic.go:334] "Generic (PLEG): container finished" podID="629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03" containerID="4fe9b07cd1599e91138683ca30e9da84b4bd93250ce15e01fd43967606252649" exitCode=0 Mar 12 13:31:52 crc kubenswrapper[4778]: I0312 13:31:52.483263 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03","Type":"ContainerDied","Data":"4fe9b07cd1599e91138683ca30e9da84b4bd93250ce15e01fd43967606252649"} Mar 12 13:31:54 crc kubenswrapper[4778]: I0312 13:31:54.791013 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4wct6" podUID="3b8efd1e-884d-4963-b69f-04ede0a92267" containerName="ovn-controller" probeResult="failure" output=< Mar 12 13:31:54 crc kubenswrapper[4778]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 12 13:31:54 crc kubenswrapper[4778]: > Mar 12 13:31:55 crc kubenswrapper[4778]: I0312 13:31:55.180819 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:55 crc kubenswrapper[4778]: I0312 13:31:55.192882 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01f943c-e09c-4727-8cf7-eec58a56b363-etc-swift\") pod \"swift-storage-0\" (UID: \"c01f943c-e09c-4727-8cf7-eec58a56b363\") " pod="openstack/swift-storage-0" Mar 12 13:31:55 crc kubenswrapper[4778]: I0312 13:31:55.268078 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 12 13:31:56 crc kubenswrapper[4778]: E0312 13:31:56.599133 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Mar 12 13:31:56 crc kubenswrapper[4778]: E0312 13:31:56.599649 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crgvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-xg6z4_openstack(befeb973-a1de-48f9-8de0-5559f75472dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:31:56 crc kubenswrapper[4778]: E0312 13:31:56.600968 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-xg6z4" podUID="befeb973-a1de-48f9-8de0-5559f75472dc" Mar 12 13:31:57 crc kubenswrapper[4778]: W0312 13:31:57.134452 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20836760_c025_4e65_bf24_34fc17f3c649.slice/crio-17a996fa52ce91166da093384fbf949661bd1de6ff1a9aade6af20fcb9e834ba WatchSource:0}: Error finding container 17a996fa52ce91166da093384fbf949661bd1de6ff1a9aade6af20fcb9e834ba: Status 404 returned error can't find the container with id 17a996fa52ce91166da093384fbf949661bd1de6ff1a9aade6af20fcb9e834ba Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.147654 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4wct6-config-6fwv4"] Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.161441 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 12 13:31:57 crc kubenswrapper[4778]: W0312 13:31:57.163659 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc01f943c_e09c_4727_8cf7_eec58a56b363.slice/crio-0697d96e46b4a903431abbb68854c041327e624887e1614c134f5e36c81aaabd WatchSource:0}: Error finding container 0697d96e46b4a903431abbb68854c041327e624887e1614c134f5e36c81aaabd: Status 404 returned error can't find the container with id 0697d96e46b4a903431abbb68854c041327e624887e1614c134f5e36c81aaabd Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.529313 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03","Type":"ContainerStarted","Data":"69b48fd30717bc2fb32adf7bd553ace0508be9fb5806d39e1e63b2dff302e279"} Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.529674 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.531325 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1e89dfcc-2ac3-444c-91e8-56991eae096b","Type":"ContainerStarted","Data":"94375f4a2a4703567f32833cc33058ee531a8b1219141eaf7fa5f176bf09075b"} Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.531523 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.532818 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"0697d96e46b4a903431abbb68854c041327e624887e1614c134f5e36c81aaabd"} Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.534598 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4wct6-config-6fwv4" event={"ID":"20836760-c025-4e65-bf24-34fc17f3c649","Type":"ContainerStarted","Data":"f28e6c324c83dfc76a63ecc641dd7e634a485b3faa88e8b16a2e55fc0961b3a8"} Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.534634 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4wct6-config-6fwv4" event={"ID":"20836760-c025-4e65-bf24-34fc17f3c649","Type":"ContainerStarted","Data":"17a996fa52ce91166da093384fbf949661bd1de6ff1a9aade6af20fcb9e834ba"} Mar 12 13:31:57 crc kubenswrapper[4778]: E0312 13:31:57.535861 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-xg6z4" podUID="befeb973-a1de-48f9-8de0-5559f75472dc" Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.559028 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371963.295773 podStartE2EDuration="1m13.559002997s" podCreationTimestamp="2026-03-12 13:30:44 +0000 UTC" firstStartedPulling="2026-03-12 13:30:46.696662484 +0000 UTC m=+1265.145357870" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:57.55595695 +0000 UTC m=+1336.004652366" watchObservedRunningTime="2026-03-12 13:31:57.559002997 +0000 UTC m=+1336.007698393" Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.582036 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4wct6-config-6fwv4" podStartSLOduration=7.582017931 podStartE2EDuration="7.582017931s" podCreationTimestamp="2026-03-12 13:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:31:57.574531418 +0000 UTC m=+1336.023226814" watchObservedRunningTime="2026-03-12 13:31:57.582017931 +0000 UTC m=+1336.030713327" Mar 12 13:31:57 crc kubenswrapper[4778]: I0312 13:31:57.635239 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=45.083803531 podStartE2EDuration="1m13.635220294s" podCreationTimestamp="2026-03-12 13:30:44 +0000 UTC" firstStartedPulling="2026-03-12 13:30:47.191925946 +0000 UTC m=+1265.640621342" lastFinishedPulling="2026-03-12 13:31:15.743342699 +0000 UTC m=+1294.192038105" observedRunningTime="2026-03-12 13:31:57.626612959 +0000 UTC m=+1336.075308355" watchObservedRunningTime="2026-03-12 13:31:57.635220294 +0000 UTC m=+1336.083915690" Mar 12 13:31:58 crc kubenswrapper[4778]: I0312 13:31:58.548814 4778 generic.go:334] "Generic (PLEG): container finished" podID="20836760-c025-4e65-bf24-34fc17f3c649" containerID="f28e6c324c83dfc76a63ecc641dd7e634a485b3faa88e8b16a2e55fc0961b3a8" exitCode=0 Mar 12 13:31:58 crc kubenswrapper[4778]: I0312 13:31:58.549111 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4wct6-config-6fwv4" event={"ID":"20836760-c025-4e65-bf24-34fc17f3c649","Type":"ContainerDied","Data":"f28e6c324c83dfc76a63ecc641dd7e634a485b3faa88e8b16a2e55fc0961b3a8"} Mar 12 13:31:59 crc kubenswrapper[4778]: I0312 13:31:59.558304 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"e51fda036eef5819395ec6d3fd22dab55d18db138a7f80e940b67e8bd1c9c47e"} Mar 12 13:31:59 crc kubenswrapper[4778]: I0312 13:31:59.558347 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"351be818ddd688aa88eb9700af5b628218d060ad816d893916ac63f3c5333bda"} Mar 12 13:31:59 crc kubenswrapper[4778]: I0312 13:31:59.558359 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"4599c0849140ef045537c5194ed1a3292aa5e8691910d4d423b2a3c0ad5ec9eb"} Mar 12 13:31:59 crc kubenswrapper[4778]: I0312 13:31:59.558369 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"b32340b64ad970ce89b19270074c18e85d0cbf22f3d3f70d70250fdd561c2684"} Mar 12 13:31:59 crc kubenswrapper[4778]: I0312 13:31:59.830940 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4wct6" Mar 12 13:31:59 crc kubenswrapper[4778]: I0312 13:31:59.913851 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.084595 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t2nc\" (UniqueName: \"kubernetes.io/projected/20836760-c025-4e65-bf24-34fc17f3c649-kube-api-access-4t2nc\") pod \"20836760-c025-4e65-bf24-34fc17f3c649\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.084754 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-log-ovn\") pod \"20836760-c025-4e65-bf24-34fc17f3c649\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.084843 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-scripts\") pod \"20836760-c025-4e65-bf24-34fc17f3c649\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.084875 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-additional-scripts\") pod \"20836760-c025-4e65-bf24-34fc17f3c649\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.084916 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run\") pod \"20836760-c025-4e65-bf24-34fc17f3c649\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.084936 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run-ovn\") pod \"20836760-c025-4e65-bf24-34fc17f3c649\" (UID: \"20836760-c025-4e65-bf24-34fc17f3c649\") " Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.085363 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "20836760-c025-4e65-bf24-34fc17f3c649" (UID: "20836760-c025-4e65-bf24-34fc17f3c649"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.086399 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "20836760-c025-4e65-bf24-34fc17f3c649" (UID: "20836760-c025-4e65-bf24-34fc17f3c649"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.086436 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run" (OuterVolumeSpecName: "var-run") pod "20836760-c025-4e65-bf24-34fc17f3c649" (UID: "20836760-c025-4e65-bf24-34fc17f3c649"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.087020 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "20836760-c025-4e65-bf24-34fc17f3c649" (UID: "20836760-c025-4e65-bf24-34fc17f3c649"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.087303 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-scripts" (OuterVolumeSpecName: "scripts") pod "20836760-c025-4e65-bf24-34fc17f3c649" (UID: "20836760-c025-4e65-bf24-34fc17f3c649"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.091429 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20836760-c025-4e65-bf24-34fc17f3c649-kube-api-access-4t2nc" (OuterVolumeSpecName: "kube-api-access-4t2nc") pod "20836760-c025-4e65-bf24-34fc17f3c649" (UID: "20836760-c025-4e65-bf24-34fc17f3c649"). InnerVolumeSpecName "kube-api-access-4t2nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.132489 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555372-rddbg"] Mar 12 13:32:00 crc kubenswrapper[4778]: E0312 13:32:00.132924 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20836760-c025-4e65-bf24-34fc17f3c649" containerName="ovn-config" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.132948 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="20836760-c025-4e65-bf24-34fc17f3c649" containerName="ovn-config" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.133142 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="20836760-c025-4e65-bf24-34fc17f3c649" containerName="ovn-config" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.133781 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.136666 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.136831 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.137463 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.147598 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555372-rddbg"] Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.187308 4778 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.187656 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.187670 4778 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/20836760-c025-4e65-bf24-34fc17f3c649-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.187686 4778 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.187698 4778 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20836760-c025-4e65-bf24-34fc17f3c649-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.187709 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t2nc\" (UniqueName: \"kubernetes.io/projected/20836760-c025-4e65-bf24-34fc17f3c649-kube-api-access-4t2nc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.216429 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4wct6-config-6fwv4"] Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.230390 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4wct6-config-6fwv4"] Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.265395 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20836760-c025-4e65-bf24-34fc17f3c649" path="/var/lib/kubelet/pods/20836760-c025-4e65-bf24-34fc17f3c649/volumes" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.289765 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl9zr\" (UniqueName: \"kubernetes.io/projected/c0b7e295-a151-42b0-a8d6-d062d9a42e88-kube-api-access-cl9zr\") pod \"auto-csr-approver-29555372-rddbg\" (UID: \"c0b7e295-a151-42b0-a8d6-d062d9a42e88\") " pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.391920 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl9zr\" (UniqueName: \"kubernetes.io/projected/c0b7e295-a151-42b0-a8d6-d062d9a42e88-kube-api-access-cl9zr\") pod \"auto-csr-approver-29555372-rddbg\" (UID: \"c0b7e295-a151-42b0-a8d6-d062d9a42e88\") " pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.429840 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl9zr\" (UniqueName: \"kubernetes.io/projected/c0b7e295-a151-42b0-a8d6-d062d9a42e88-kube-api-access-cl9zr\") pod \"auto-csr-approver-29555372-rddbg\" (UID: \"c0b7e295-a151-42b0-a8d6-d062d9a42e88\") " pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.455857 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.578725 4778 scope.go:117] "RemoveContainer" containerID="f28e6c324c83dfc76a63ecc641dd7e634a485b3faa88e8b16a2e55fc0961b3a8" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.578791 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4wct6-config-6fwv4" Mar 12 13:32:00 crc kubenswrapper[4778]: I0312 13:32:00.894753 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555372-rddbg"] Mar 12 13:32:01 crc kubenswrapper[4778]: I0312 13:32:01.591799 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"f68fe853c5f2eff9d19ab2850f709aa98efa3ecb7134c5c1e61c852033409890"} Mar 12 13:32:01 crc kubenswrapper[4778]: I0312 13:32:01.592043 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"f236f31858d7f8903c2548d01c7eb7e051fb9c7e772aea262b1dc82f06888421"} Mar 12 13:32:01 crc kubenswrapper[4778]: I0312 13:32:01.592054 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"4a599b99dd2b156167a96adf804cb32cee43a8a4c720de16059ea78b2fd6e31c"} Mar 12 13:32:01 crc kubenswrapper[4778]: I0312 13:32:01.592065 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"f28d138f1fd7bc3661dfeeb07a61f60a94fa3f6a200c91f560af4ae4c400dd78"} Mar 12 13:32:01 crc kubenswrapper[4778]: I0312 13:32:01.595413 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555372-rddbg" event={"ID":"c0b7e295-a151-42b0-a8d6-d062d9a42e88","Type":"ContainerStarted","Data":"b90b0c8b3d3e98566eca6281f35c9f6e9b84e5323c00deb50365e0df5d3b91e3"} Mar 12 13:32:02 crc kubenswrapper[4778]: I0312 13:32:02.602994 4778 generic.go:334] "Generic (PLEG): container finished" podID="c0b7e295-a151-42b0-a8d6-d062d9a42e88" containerID="83e30e12aea92ff26adeced3b96dea20e98c42e4bd6fda29118e167bf1eeb711" exitCode=0 Mar 12 13:32:02 crc kubenswrapper[4778]: I0312 13:32:02.603050 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555372-rddbg" event={"ID":"c0b7e295-a151-42b0-a8d6-d062d9a42e88","Type":"ContainerDied","Data":"83e30e12aea92ff26adeced3b96dea20e98c42e4bd6fda29118e167bf1eeb711"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.626520 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"71c934b93dc81fcb4c6490ad1cb01c2110aadcacc2c2b6331fb1f63216515fae"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.626602 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"d537ea6c9c04cfb9e576ff7eb8061b47046893f4ff7b34df3bb0907f800a376a"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.626620 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"0b84c79df75a4a2e3a8aebea9c2fd3414a940b4674e32f45e81b08a68fe8d46f"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.626633 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"e15d95880cc0b746ab57af968115fecac856e5aedb268b4dae48ba5038796103"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.626667 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"908dae9220f9ffa18b60246e05ac363811858baf5deb16715e04c852b4af38f9"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.626680 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"a41cbc225bec1ba427df8fdd647922f8cb16681229c182bed96fec2cb17265b5"} Mar 12 13:32:03 crc kubenswrapper[4778]: I0312 13:32:03.933011 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.053702 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl9zr\" (UniqueName: \"kubernetes.io/projected/c0b7e295-a151-42b0-a8d6-d062d9a42e88-kube-api-access-cl9zr\") pod \"c0b7e295-a151-42b0-a8d6-d062d9a42e88\" (UID: \"c0b7e295-a151-42b0-a8d6-d062d9a42e88\") " Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.059895 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b7e295-a151-42b0-a8d6-d062d9a42e88-kube-api-access-cl9zr" (OuterVolumeSpecName: "kube-api-access-cl9zr") pod "c0b7e295-a151-42b0-a8d6-d062d9a42e88" (UID: "c0b7e295-a151-42b0-a8d6-d062d9a42e88"). InnerVolumeSpecName "kube-api-access-cl9zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.155916 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl9zr\" (UniqueName: \"kubernetes.io/projected/c0b7e295-a151-42b0-a8d6-d062d9a42e88-kube-api-access-cl9zr\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.648705 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555372-rddbg" event={"ID":"c0b7e295-a151-42b0-a8d6-d062d9a42e88","Type":"ContainerDied","Data":"b90b0c8b3d3e98566eca6281f35c9f6e9b84e5323c00deb50365e0df5d3b91e3"} Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.648958 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b90b0c8b3d3e98566eca6281f35c9f6e9b84e5323c00deb50365e0df5d3b91e3" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.649019 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555372-rddbg" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.658421 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c01f943c-e09c-4727-8cf7-eec58a56b363","Type":"ContainerStarted","Data":"90a3afd022e6450b75eb74b1a883ba0a851eec230b95297ba3e7d6063ef22c6a"} Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.713605 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.351960756 podStartE2EDuration="42.713584985s" podCreationTimestamp="2026-03-12 13:31:22 +0000 UTC" firstStartedPulling="2026-03-12 13:31:57.166338432 +0000 UTC m=+1335.615033838" lastFinishedPulling="2026-03-12 13:32:02.527962671 +0000 UTC m=+1340.976658067" observedRunningTime="2026-03-12 13:32:04.70813717 +0000 UTC m=+1343.156832576" watchObservedRunningTime="2026-03-12 13:32:04.713584985 +0000 UTC m=+1343.162280391" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.985008 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-56bl9"] Mar 12 13:32:04 crc kubenswrapper[4778]: E0312 13:32:04.985426 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b7e295-a151-42b0-a8d6-d062d9a42e88" containerName="oc" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.985446 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b7e295-a151-42b0-a8d6-d062d9a42e88" containerName="oc" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.985636 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b7e295-a151-42b0-a8d6-d062d9a42e88" containerName="oc" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.986715 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:04 crc kubenswrapper[4778]: I0312 13:32:04.988486 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.000326 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-56bl9"] Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.050247 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555366-zt5bk"] Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073100 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555366-zt5bk"] Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073488 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-svc\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073564 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073643 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-config\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073667 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073688 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsg69\" (UniqueName: \"kubernetes.io/projected/811bc15c-050c-4d37-a19f-095086748286-kube-api-access-bsg69\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.073721 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.175084 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.175257 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-config\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.175288 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.175322 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsg69\" (UniqueName: \"kubernetes.io/projected/811bc15c-050c-4d37-a19f-095086748286-kube-api-access-bsg69\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.175370 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.175440 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-svc\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.176225 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.176275 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-config\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.176294 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.176515 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.176559 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-svc\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.195008 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsg69\" (UniqueName: \"kubernetes.io/projected/811bc15c-050c-4d37-a19f-095086748286-kube-api-access-bsg69\") pod \"dnsmasq-dns-764c5664d7-56bl9\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.304702 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:05 crc kubenswrapper[4778]: I0312 13:32:05.780779 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-56bl9"] Mar 12 13:32:05 crc kubenswrapper[4778]: W0312 13:32:05.791237 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod811bc15c_050c_4d37_a19f_095086748286.slice/crio-9e0eacf82432587cd58359c3985b8def0ae32125ba66b4e86532ed5c793bbd04 WatchSource:0}: Error finding container 9e0eacf82432587cd58359c3985b8def0ae32125ba66b4e86532ed5c793bbd04: Status 404 returned error can't find the container with id 9e0eacf82432587cd58359c3985b8def0ae32125ba66b4e86532ed5c793bbd04 Mar 12 13:32:06 crc kubenswrapper[4778]: I0312 13:32:06.039371 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 12 13:32:06 crc kubenswrapper[4778]: I0312 13:32:06.267694 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d48c598c-314b-4dc6-af90-7772a2ca7f2d" path="/var/lib/kubelet/pods/d48c598c-314b-4dc6-af90-7772a2ca7f2d/volumes" Mar 12 13:32:06 crc kubenswrapper[4778]: I0312 13:32:06.672761 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 12 13:32:06 crc kubenswrapper[4778]: I0312 13:32:06.675620 4778 generic.go:334] "Generic (PLEG): container finished" podID="811bc15c-050c-4d37-a19f-095086748286" containerID="52a29e484c375a20ac3f8fc8c2aa037eb3038bed507119d164be5bd117815abc" exitCode=0 Mar 12 13:32:06 crc kubenswrapper[4778]: I0312 13:32:06.675667 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" event={"ID":"811bc15c-050c-4d37-a19f-095086748286","Type":"ContainerDied","Data":"52a29e484c375a20ac3f8fc8c2aa037eb3038bed507119d164be5bd117815abc"} Mar 12 13:32:06 crc kubenswrapper[4778]: I0312 13:32:06.675696 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" event={"ID":"811bc15c-050c-4d37-a19f-095086748286","Type":"ContainerStarted","Data":"9e0eacf82432587cd58359c3985b8def0ae32125ba66b4e86532ed5c793bbd04"} Mar 12 13:32:07 crc kubenswrapper[4778]: I0312 13:32:07.688524 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" event={"ID":"811bc15c-050c-4d37-a19f-095086748286","Type":"ContainerStarted","Data":"512c2c0cf187f0ee46cccf1da3f29d083846818126627409ab7b1bb5fa1ef052"} Mar 12 13:32:07 crc kubenswrapper[4778]: I0312 13:32:07.688889 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:07 crc kubenswrapper[4778]: I0312 13:32:07.710602 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" podStartSLOduration=3.7105742 podStartE2EDuration="3.7105742s" podCreationTimestamp="2026-03-12 13:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:07.70740525 +0000 UTC m=+1346.156100656" watchObservedRunningTime="2026-03-12 13:32:07.7105742 +0000 UTC m=+1346.159269606" Mar 12 13:32:07 crc kubenswrapper[4778]: I0312 13:32:07.979402 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-nh9xs"] Mar 12 13:32:07 crc kubenswrapper[4778]: I0312 13:32:07.980613 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:07 crc kubenswrapper[4778]: I0312 13:32:07.993516 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nh9xs"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.134565 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdrjg\" (UniqueName: \"kubernetes.io/projected/3800be73-3a09-42b6-8d01-592ccbc6aaa3-kube-api-access-tdrjg\") pod \"cinder-db-create-nh9xs\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.134648 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3800be73-3a09-42b6-8d01-592ccbc6aaa3-operator-scripts\") pod \"cinder-db-create-nh9xs\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.192714 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-31ed-account-create-update-h8bhm"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.193939 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.196650 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.205764 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-31ed-account-create-update-h8bhm"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.235960 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdrjg\" (UniqueName: \"kubernetes.io/projected/3800be73-3a09-42b6-8d01-592ccbc6aaa3-kube-api-access-tdrjg\") pod \"cinder-db-create-nh9xs\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.236047 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3800be73-3a09-42b6-8d01-592ccbc6aaa3-operator-scripts\") pod \"cinder-db-create-nh9xs\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.236766 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3800be73-3a09-42b6-8d01-592ccbc6aaa3-operator-scripts\") pod \"cinder-db-create-nh9xs\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.254265 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdrjg\" (UniqueName: \"kubernetes.io/projected/3800be73-3a09-42b6-8d01-592ccbc6aaa3-kube-api-access-tdrjg\") pod \"cinder-db-create-nh9xs\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.288776 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-gxsm6"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.293727 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.306981 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gxsm6"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.337132 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsvdw\" (UniqueName: \"kubernetes.io/projected/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-kube-api-access-vsvdw\") pod \"cinder-31ed-account-create-update-h8bhm\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.337298 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-operator-scripts\") pod \"cinder-31ed-account-create-update-h8bhm\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.340513 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.388813 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2abd-account-create-update-chtfz"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.389887 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.391127 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.408324 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2abd-account-create-update-chtfz"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.438660 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsvdw\" (UniqueName: \"kubernetes.io/projected/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-kube-api-access-vsvdw\") pod \"cinder-31ed-account-create-update-h8bhm\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.438750 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnm5n\" (UniqueName: \"kubernetes.io/projected/79ff3988-1976-4049-8277-0acb36da44c5-kube-api-access-rnm5n\") pod \"barbican-db-create-gxsm6\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.438863 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-operator-scripts\") pod \"cinder-31ed-account-create-update-h8bhm\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.438937 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79ff3988-1976-4049-8277-0acb36da44c5-operator-scripts\") pod \"barbican-db-create-gxsm6\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.439783 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-operator-scripts\") pod \"cinder-31ed-account-create-update-h8bhm\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.451534 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-57cfm"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.452492 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.471285 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.471704 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.471861 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rjpsk" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.472047 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.487350 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-57cfm"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.496706 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsvdw\" (UniqueName: \"kubernetes.io/projected/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-kube-api-access-vsvdw\") pod \"cinder-31ed-account-create-update-h8bhm\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.510930 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.519561 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e25a-account-create-update-vs6zm"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.520525 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.530121 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.540582 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e25a-account-create-update-vs6zm"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.553381 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnm5n\" (UniqueName: \"kubernetes.io/projected/79ff3988-1976-4049-8277-0acb36da44c5-kube-api-access-rnm5n\") pod \"barbican-db-create-gxsm6\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.553739 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79ff3988-1976-4049-8277-0acb36da44c5-operator-scripts\") pod \"barbican-db-create-gxsm6\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.553839 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-config-data\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.553928 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfc2k\" (UniqueName: \"kubernetes.io/projected/729468a8-fded-4564-96c8-471d3cf48825-kube-api-access-tfc2k\") pod \"neutron-2abd-account-create-update-chtfz\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.554013 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/729468a8-fded-4564-96c8-471d3cf48825-operator-scripts\") pod \"neutron-2abd-account-create-update-chtfz\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.554107 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-combined-ca-bundle\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.554176 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlp2n\" (UniqueName: \"kubernetes.io/projected/ec77eae6-4dac-4535-b0d3-98bd3422e4de-kube-api-access-xlp2n\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.554977 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79ff3988-1976-4049-8277-0acb36da44c5-operator-scripts\") pod \"barbican-db-create-gxsm6\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.576966 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnm5n\" (UniqueName: \"kubernetes.io/projected/79ff3988-1976-4049-8277-0acb36da44c5-kube-api-access-rnm5n\") pod \"barbican-db-create-gxsm6\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.589474 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-thsh7"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.594113 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.610444 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-thsh7"] Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.614121 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655612 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31f8bb53-a8a8-448f-8f42-349232e383ec-operator-scripts\") pod \"barbican-e25a-account-create-update-vs6zm\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655742 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-config-data\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655779 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfc2k\" (UniqueName: \"kubernetes.io/projected/729468a8-fded-4564-96c8-471d3cf48825-kube-api-access-tfc2k\") pod \"neutron-2abd-account-create-update-chtfz\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655807 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/729468a8-fded-4564-96c8-471d3cf48825-operator-scripts\") pod \"neutron-2abd-account-create-update-chtfz\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655841 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-operator-scripts\") pod \"neutron-db-create-thsh7\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655878 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-combined-ca-bundle\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.655904 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlp2n\" (UniqueName: \"kubernetes.io/projected/ec77eae6-4dac-4535-b0d3-98bd3422e4de-kube-api-access-xlp2n\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.656030 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtxwj\" (UniqueName: \"kubernetes.io/projected/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-kube-api-access-dtxwj\") pod \"neutron-db-create-thsh7\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.656061 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbpjb\" (UniqueName: \"kubernetes.io/projected/31f8bb53-a8a8-448f-8f42-349232e383ec-kube-api-access-wbpjb\") pod \"barbican-e25a-account-create-update-vs6zm\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.659505 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/729468a8-fded-4564-96c8-471d3cf48825-operator-scripts\") pod \"neutron-2abd-account-create-update-chtfz\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.660279 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-config-data\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.663489 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-combined-ca-bundle\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.678958 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfc2k\" (UniqueName: \"kubernetes.io/projected/729468a8-fded-4564-96c8-471d3cf48825-kube-api-access-tfc2k\") pod \"neutron-2abd-account-create-update-chtfz\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.685568 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlp2n\" (UniqueName: \"kubernetes.io/projected/ec77eae6-4dac-4535-b0d3-98bd3422e4de-kube-api-access-xlp2n\") pod \"keystone-db-sync-57cfm\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.744887 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.758101 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtxwj\" (UniqueName: \"kubernetes.io/projected/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-kube-api-access-dtxwj\") pod \"neutron-db-create-thsh7\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.758139 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbpjb\" (UniqueName: \"kubernetes.io/projected/31f8bb53-a8a8-448f-8f42-349232e383ec-kube-api-access-wbpjb\") pod \"barbican-e25a-account-create-update-vs6zm\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.758278 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31f8bb53-a8a8-448f-8f42-349232e383ec-operator-scripts\") pod \"barbican-e25a-account-create-update-vs6zm\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.758341 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-operator-scripts\") pod \"neutron-db-create-thsh7\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.759251 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31f8bb53-a8a8-448f-8f42-349232e383ec-operator-scripts\") pod \"barbican-e25a-account-create-update-vs6zm\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.762230 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-operator-scripts\") pod \"neutron-db-create-thsh7\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.775170 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbpjb\" (UniqueName: \"kubernetes.io/projected/31f8bb53-a8a8-448f-8f42-349232e383ec-kube-api-access-wbpjb\") pod \"barbican-e25a-account-create-update-vs6zm\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.778742 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtxwj\" (UniqueName: \"kubernetes.io/projected/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-kube-api-access-dtxwj\") pod \"neutron-db-create-thsh7\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.815854 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.896435 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.922233 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:08 crc kubenswrapper[4778]: I0312 13:32:08.947856 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nh9xs"] Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.030887 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-31ed-account-create-update-h8bhm"] Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.166428 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-gxsm6"] Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.301474 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2abd-account-create-update-chtfz"] Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.413569 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-57cfm"] Mar 12 13:32:09 crc kubenswrapper[4778]: W0312 13:32:09.417521 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec77eae6_4dac_4535_b0d3_98bd3422e4de.slice/crio-e416653bf42b2d593f800644adfdb9dd57501c2cbfadea4a9a6bd3bc9f20d011 WatchSource:0}: Error finding container e416653bf42b2d593f800644adfdb9dd57501c2cbfadea4a9a6bd3bc9f20d011: Status 404 returned error can't find the container with id e416653bf42b2d593f800644adfdb9dd57501c2cbfadea4a9a6bd3bc9f20d011 Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.504109 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-thsh7"] Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.518830 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e25a-account-create-update-vs6zm"] Mar 12 13:32:09 crc kubenswrapper[4778]: W0312 13:32:09.544018 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31f8bb53_a8a8_448f_8f42_349232e383ec.slice/crio-7af48946578bd61f228568d72f3eb420af6ca8e366fa1043f54b2e44d5cf2462 WatchSource:0}: Error finding container 7af48946578bd61f228568d72f3eb420af6ca8e366fa1043f54b2e44d5cf2462: Status 404 returned error can't find the container with id 7af48946578bd61f228568d72f3eb420af6ca8e366fa1043f54b2e44d5cf2462 Mar 12 13:32:09 crc kubenswrapper[4778]: W0312 13:32:09.544511 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9793dfb5_c2a5_4dc1_993d_9e024a810ce8.slice/crio-f4df7f9c0ba0bb575b3f8143a0a2a4366ed16f5aa6229ee4e07f76139f3bdb21 WatchSource:0}: Error finding container f4df7f9c0ba0bb575b3f8143a0a2a4366ed16f5aa6229ee4e07f76139f3bdb21: Status 404 returned error can't find the container with id f4df7f9c0ba0bb575b3f8143a0a2a4366ed16f5aa6229ee4e07f76139f3bdb21 Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.719202 4778 generic.go:334] "Generic (PLEG): container finished" podID="79ff3988-1976-4049-8277-0acb36da44c5" containerID="93602c5ae72cfd4f9a42c4921524905037c8077ce8260918d72d9601b072dd59" exitCode=0 Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.719327 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gxsm6" event={"ID":"79ff3988-1976-4049-8277-0acb36da44c5","Type":"ContainerDied","Data":"93602c5ae72cfd4f9a42c4921524905037c8077ce8260918d72d9601b072dd59"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.719378 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gxsm6" event={"ID":"79ff3988-1976-4049-8277-0acb36da44c5","Type":"ContainerStarted","Data":"8a4949a97a8077a1a4d6a1981effa75bacbbaf8684ec8ee1397aba5987fcdd14"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.721493 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2abd-account-create-update-chtfz" event={"ID":"729468a8-fded-4564-96c8-471d3cf48825","Type":"ContainerStarted","Data":"a62186594073bc08d5194d8b9ce9a46d1a29b359b5ca56b7c0f8fed38f9c7470"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.721526 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2abd-account-create-update-chtfz" event={"ID":"729468a8-fded-4564-96c8-471d3cf48825","Type":"ContainerStarted","Data":"9dc728b0f4cc0eafb30fa27920b49b62f951fc64e849dafb2097b41e077244e3"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.726661 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-thsh7" event={"ID":"9793dfb5-c2a5-4dc1-993d-9e024a810ce8","Type":"ContainerStarted","Data":"3bf3addaa75cf85838ea1739e9760ca68c0ed5921fd1bd5da9e4725715df9a99"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.726713 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-thsh7" event={"ID":"9793dfb5-c2a5-4dc1-993d-9e024a810ce8","Type":"ContainerStarted","Data":"f4df7f9c0ba0bb575b3f8143a0a2a4366ed16f5aa6229ee4e07f76139f3bdb21"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.728829 4778 generic.go:334] "Generic (PLEG): container finished" podID="3800be73-3a09-42b6-8d01-592ccbc6aaa3" containerID="70fc2c631648b6cf05ce7c564c8a25d897ce94ea350c4d6a8a0ccacb6c5f16b4" exitCode=0 Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.728887 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nh9xs" event={"ID":"3800be73-3a09-42b6-8d01-592ccbc6aaa3","Type":"ContainerDied","Data":"70fc2c631648b6cf05ce7c564c8a25d897ce94ea350c4d6a8a0ccacb6c5f16b4"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.728910 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nh9xs" event={"ID":"3800be73-3a09-42b6-8d01-592ccbc6aaa3","Type":"ContainerStarted","Data":"95d2ed0897cf2bbe34932a1f54f1dbbea9a78e7b987d5a18fb616fc0888d29bf"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.729971 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57cfm" event={"ID":"ec77eae6-4dac-4535-b0d3-98bd3422e4de","Type":"ContainerStarted","Data":"e416653bf42b2d593f800644adfdb9dd57501c2cbfadea4a9a6bd3bc9f20d011"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.733171 4778 generic.go:334] "Generic (PLEG): container finished" podID="4b694c81-3b07-45a1-9ca1-1e47e7430f1f" containerID="13ffa46dd0ede6f8f4fd6e787f1d2948d8a5e96a8e47df52e40147817681f0f7" exitCode=0 Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.733233 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-31ed-account-create-update-h8bhm" event={"ID":"4b694c81-3b07-45a1-9ca1-1e47e7430f1f","Type":"ContainerDied","Data":"13ffa46dd0ede6f8f4fd6e787f1d2948d8a5e96a8e47df52e40147817681f0f7"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.733372 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-31ed-account-create-update-h8bhm" event={"ID":"4b694c81-3b07-45a1-9ca1-1e47e7430f1f","Type":"ContainerStarted","Data":"f7445a7703568cb1d2b72115b07ec7b8761dc64a43181a517180c2f025e9efc9"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.734837 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e25a-account-create-update-vs6zm" event={"ID":"31f8bb53-a8a8-448f-8f42-349232e383ec","Type":"ContainerStarted","Data":"7af48946578bd61f228568d72f3eb420af6ca8e366fa1043f54b2e44d5cf2462"} Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.761131 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-thsh7" podStartSLOduration=1.761109894 podStartE2EDuration="1.761109894s" podCreationTimestamp="2026-03-12 13:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:09.752471688 +0000 UTC m=+1348.201167084" watchObservedRunningTime="2026-03-12 13:32:09.761109894 +0000 UTC m=+1348.209805290" Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.777018 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-2abd-account-create-update-chtfz" podStartSLOduration=1.777001316 podStartE2EDuration="1.777001316s" podCreationTimestamp="2026-03-12 13:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:09.773487696 +0000 UTC m=+1348.222183092" watchObservedRunningTime="2026-03-12 13:32:09.777001316 +0000 UTC m=+1348.225696702" Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.795354 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-e25a-account-create-update-vs6zm" podStartSLOduration=1.795334947 podStartE2EDuration="1.795334947s" podCreationTimestamp="2026-03-12 13:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:09.789544742 +0000 UTC m=+1348.238240138" watchObservedRunningTime="2026-03-12 13:32:09.795334947 +0000 UTC m=+1348.244030343" Mar 12 13:32:09 crc kubenswrapper[4778]: I0312 13:32:09.860213 4778 scope.go:117] "RemoveContainer" containerID="59816c72d24ee82ad1e212a580fdeb3c8cd671c1f79b421c31d995678ebec873" Mar 12 13:32:10 crc kubenswrapper[4778]: I0312 13:32:10.747855 4778 generic.go:334] "Generic (PLEG): container finished" podID="31f8bb53-a8a8-448f-8f42-349232e383ec" containerID="60b2242b65665faad21e5afc28edb1788f01dc784524abe26ac1b4cb9a5296a5" exitCode=0 Mar 12 13:32:10 crc kubenswrapper[4778]: I0312 13:32:10.747946 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e25a-account-create-update-vs6zm" event={"ID":"31f8bb53-a8a8-448f-8f42-349232e383ec","Type":"ContainerDied","Data":"60b2242b65665faad21e5afc28edb1788f01dc784524abe26ac1b4cb9a5296a5"} Mar 12 13:32:10 crc kubenswrapper[4778]: I0312 13:32:10.749997 4778 generic.go:334] "Generic (PLEG): container finished" podID="729468a8-fded-4564-96c8-471d3cf48825" containerID="a62186594073bc08d5194d8b9ce9a46d1a29b359b5ca56b7c0f8fed38f9c7470" exitCode=0 Mar 12 13:32:10 crc kubenswrapper[4778]: I0312 13:32:10.750096 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2abd-account-create-update-chtfz" event={"ID":"729468a8-fded-4564-96c8-471d3cf48825","Type":"ContainerDied","Data":"a62186594073bc08d5194d8b9ce9a46d1a29b359b5ca56b7c0f8fed38f9c7470"} Mar 12 13:32:10 crc kubenswrapper[4778]: I0312 13:32:10.751781 4778 generic.go:334] "Generic (PLEG): container finished" podID="9793dfb5-c2a5-4dc1-993d-9e024a810ce8" containerID="3bf3addaa75cf85838ea1739e9760ca68c0ed5921fd1bd5da9e4725715df9a99" exitCode=0 Mar 12 13:32:10 crc kubenswrapper[4778]: I0312 13:32:10.751822 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-thsh7" event={"ID":"9793dfb5-c2a5-4dc1-993d-9e024a810ce8","Type":"ContainerDied","Data":"3bf3addaa75cf85838ea1739e9760ca68c0ed5921fd1bd5da9e4725715df9a99"} Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.089589 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.184570 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.187815 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.232252 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdrjg\" (UniqueName: \"kubernetes.io/projected/3800be73-3a09-42b6-8d01-592ccbc6aaa3-kube-api-access-tdrjg\") pod \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.232385 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3800be73-3a09-42b6-8d01-592ccbc6aaa3-operator-scripts\") pod \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\" (UID: \"3800be73-3a09-42b6-8d01-592ccbc6aaa3\") " Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.233311 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3800be73-3a09-42b6-8d01-592ccbc6aaa3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3800be73-3a09-42b6-8d01-592ccbc6aaa3" (UID: "3800be73-3a09-42b6-8d01-592ccbc6aaa3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.239937 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3800be73-3a09-42b6-8d01-592ccbc6aaa3-kube-api-access-tdrjg" (OuterVolumeSpecName: "kube-api-access-tdrjg") pod "3800be73-3a09-42b6-8d01-592ccbc6aaa3" (UID: "3800be73-3a09-42b6-8d01-592ccbc6aaa3"). InnerVolumeSpecName "kube-api-access-tdrjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.333474 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsvdw\" (UniqueName: \"kubernetes.io/projected/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-kube-api-access-vsvdw\") pod \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.333544 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnm5n\" (UniqueName: \"kubernetes.io/projected/79ff3988-1976-4049-8277-0acb36da44c5-kube-api-access-rnm5n\") pod \"79ff3988-1976-4049-8277-0acb36da44c5\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.333570 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79ff3988-1976-4049-8277-0acb36da44c5-operator-scripts\") pod \"79ff3988-1976-4049-8277-0acb36da44c5\" (UID: \"79ff3988-1976-4049-8277-0acb36da44c5\") " Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.333618 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-operator-scripts\") pod \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\" (UID: \"4b694c81-3b07-45a1-9ca1-1e47e7430f1f\") " Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.334157 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdrjg\" (UniqueName: \"kubernetes.io/projected/3800be73-3a09-42b6-8d01-592ccbc6aaa3-kube-api-access-tdrjg\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.334173 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3800be73-3a09-42b6-8d01-592ccbc6aaa3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.334619 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4b694c81-3b07-45a1-9ca1-1e47e7430f1f" (UID: "4b694c81-3b07-45a1-9ca1-1e47e7430f1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.335003 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79ff3988-1976-4049-8277-0acb36da44c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79ff3988-1976-4049-8277-0acb36da44c5" (UID: "79ff3988-1976-4049-8277-0acb36da44c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.338603 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-kube-api-access-vsvdw" (OuterVolumeSpecName: "kube-api-access-vsvdw") pod "4b694c81-3b07-45a1-9ca1-1e47e7430f1f" (UID: "4b694c81-3b07-45a1-9ca1-1e47e7430f1f"). InnerVolumeSpecName "kube-api-access-vsvdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.339122 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ff3988-1976-4049-8277-0acb36da44c5-kube-api-access-rnm5n" (OuterVolumeSpecName: "kube-api-access-rnm5n") pod "79ff3988-1976-4049-8277-0acb36da44c5" (UID: "79ff3988-1976-4049-8277-0acb36da44c5"). InnerVolumeSpecName "kube-api-access-rnm5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.436084 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsvdw\" (UniqueName: \"kubernetes.io/projected/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-kube-api-access-vsvdw\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.436118 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnm5n\" (UniqueName: \"kubernetes.io/projected/79ff3988-1976-4049-8277-0acb36da44c5-kube-api-access-rnm5n\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.436127 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79ff3988-1976-4049-8277-0acb36da44c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.436136 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b694c81-3b07-45a1-9ca1-1e47e7430f1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.776501 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nh9xs" event={"ID":"3800be73-3a09-42b6-8d01-592ccbc6aaa3","Type":"ContainerDied","Data":"95d2ed0897cf2bbe34932a1f54f1dbbea9a78e7b987d5a18fb616fc0888d29bf"} Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.776749 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95d2ed0897cf2bbe34932a1f54f1dbbea9a78e7b987d5a18fb616fc0888d29bf" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.776805 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nh9xs" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.791872 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-31ed-account-create-update-h8bhm" event={"ID":"4b694c81-3b07-45a1-9ca1-1e47e7430f1f","Type":"ContainerDied","Data":"f7445a7703568cb1d2b72115b07ec7b8761dc64a43181a517180c2f025e9efc9"} Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.791900 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7445a7703568cb1d2b72115b07ec7b8761dc64a43181a517180c2f025e9efc9" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.791960 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31ed-account-create-update-h8bhm" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.795626 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-gxsm6" event={"ID":"79ff3988-1976-4049-8277-0acb36da44c5","Type":"ContainerDied","Data":"8a4949a97a8077a1a4d6a1981effa75bacbbaf8684ec8ee1397aba5987fcdd14"} Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.795649 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a4949a97a8077a1a4d6a1981effa75bacbbaf8684ec8ee1397aba5987fcdd14" Mar 12 13:32:11 crc kubenswrapper[4778]: I0312 13:32:11.795680 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-gxsm6" Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.812531 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-thsh7" event={"ID":"9793dfb5-c2a5-4dc1-993d-9e024a810ce8","Type":"ContainerDied","Data":"f4df7f9c0ba0bb575b3f8143a0a2a4366ed16f5aa6229ee4e07f76139f3bdb21"} Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.812871 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4df7f9c0ba0bb575b3f8143a0a2a4366ed16f5aa6229ee4e07f76139f3bdb21" Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.814084 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e25a-account-create-update-vs6zm" event={"ID":"31f8bb53-a8a8-448f-8f42-349232e383ec","Type":"ContainerDied","Data":"7af48946578bd61f228568d72f3eb420af6ca8e366fa1043f54b2e44d5cf2462"} Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.814122 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7af48946578bd61f228568d72f3eb420af6ca8e366fa1043f54b2e44d5cf2462" Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.819912 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2abd-account-create-update-chtfz" event={"ID":"729468a8-fded-4564-96c8-471d3cf48825","Type":"ContainerDied","Data":"9dc728b0f4cc0eafb30fa27920b49b62f951fc64e849dafb2097b41e077244e3"} Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.819960 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dc728b0f4cc0eafb30fa27920b49b62f951fc64e849dafb2097b41e077244e3" Mar 12 13:32:13 crc kubenswrapper[4778]: I0312 13:32:13.978537 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.004143 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.016372 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.074556 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfc2k\" (UniqueName: \"kubernetes.io/projected/729468a8-fded-4564-96c8-471d3cf48825-kube-api-access-tfc2k\") pod \"729468a8-fded-4564-96c8-471d3cf48825\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.074659 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-operator-scripts\") pod \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.074685 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/729468a8-fded-4564-96c8-471d3cf48825-operator-scripts\") pod \"729468a8-fded-4564-96c8-471d3cf48825\" (UID: \"729468a8-fded-4564-96c8-471d3cf48825\") " Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.074699 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtxwj\" (UniqueName: \"kubernetes.io/projected/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-kube-api-access-dtxwj\") pod \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\" (UID: \"9793dfb5-c2a5-4dc1-993d-9e024a810ce8\") " Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.076153 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9793dfb5-c2a5-4dc1-993d-9e024a810ce8" (UID: "9793dfb5-c2a5-4dc1-993d-9e024a810ce8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.076648 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/729468a8-fded-4564-96c8-471d3cf48825-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "729468a8-fded-4564-96c8-471d3cf48825" (UID: "729468a8-fded-4564-96c8-471d3cf48825"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.079336 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/729468a8-fded-4564-96c8-471d3cf48825-kube-api-access-tfc2k" (OuterVolumeSpecName: "kube-api-access-tfc2k") pod "729468a8-fded-4564-96c8-471d3cf48825" (UID: "729468a8-fded-4564-96c8-471d3cf48825"). InnerVolumeSpecName "kube-api-access-tfc2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.079476 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-kube-api-access-dtxwj" (OuterVolumeSpecName: "kube-api-access-dtxwj") pod "9793dfb5-c2a5-4dc1-993d-9e024a810ce8" (UID: "9793dfb5-c2a5-4dc1-993d-9e024a810ce8"). InnerVolumeSpecName "kube-api-access-dtxwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.176346 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbpjb\" (UniqueName: \"kubernetes.io/projected/31f8bb53-a8a8-448f-8f42-349232e383ec-kube-api-access-wbpjb\") pod \"31f8bb53-a8a8-448f-8f42-349232e383ec\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.176432 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31f8bb53-a8a8-448f-8f42-349232e383ec-operator-scripts\") pod \"31f8bb53-a8a8-448f-8f42-349232e383ec\" (UID: \"31f8bb53-a8a8-448f-8f42-349232e383ec\") " Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.176746 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfc2k\" (UniqueName: \"kubernetes.io/projected/729468a8-fded-4564-96c8-471d3cf48825-kube-api-access-tfc2k\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.176763 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.176772 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtxwj\" (UniqueName: \"kubernetes.io/projected/9793dfb5-c2a5-4dc1-993d-9e024a810ce8-kube-api-access-dtxwj\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.176782 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/729468a8-fded-4564-96c8-471d3cf48825-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.177015 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31f8bb53-a8a8-448f-8f42-349232e383ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31f8bb53-a8a8-448f-8f42-349232e383ec" (UID: "31f8bb53-a8a8-448f-8f42-349232e383ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.186394 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f8bb53-a8a8-448f-8f42-349232e383ec-kube-api-access-wbpjb" (OuterVolumeSpecName: "kube-api-access-wbpjb") pod "31f8bb53-a8a8-448f-8f42-349232e383ec" (UID: "31f8bb53-a8a8-448f-8f42-349232e383ec"). InnerVolumeSpecName "kube-api-access-wbpjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.280460 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbpjb\" (UniqueName: \"kubernetes.io/projected/31f8bb53-a8a8-448f-8f42-349232e383ec-kube-api-access-wbpjb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.280491 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31f8bb53-a8a8-448f-8f42-349232e383ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.829793 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57cfm" event={"ID":"ec77eae6-4dac-4535-b0d3-98bd3422e4de","Type":"ContainerStarted","Data":"fc1fdc3b0586065e85920687a0b5a3f3a3005e79a719fda2a25493dca50c853e"} Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.831414 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2abd-account-create-update-chtfz" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.831494 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-thsh7" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.831417 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xg6z4" event={"ID":"befeb973-a1de-48f9-8de0-5559f75472dc","Type":"ContainerStarted","Data":"58438369e99b6009fb9ed545548de66fcc857634b3821d960d6e5735646c9d5c"} Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.831792 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e25a-account-create-update-vs6zm" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.849076 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-57cfm" podStartSLOduration=2.447329164 podStartE2EDuration="6.8490285s" podCreationTimestamp="2026-03-12 13:32:08 +0000 UTC" firstStartedPulling="2026-03-12 13:32:09.419702296 +0000 UTC m=+1347.868397692" lastFinishedPulling="2026-03-12 13:32:13.821401622 +0000 UTC m=+1352.270097028" observedRunningTime="2026-03-12 13:32:14.847540358 +0000 UTC m=+1353.296235744" watchObservedRunningTime="2026-03-12 13:32:14.8490285 +0000 UTC m=+1353.297723916" Mar 12 13:32:14 crc kubenswrapper[4778]: I0312 13:32:14.863625 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-xg6z4" podStartSLOduration=2.7914096710000003 podStartE2EDuration="35.863602524s" podCreationTimestamp="2026-03-12 13:31:39 +0000 UTC" firstStartedPulling="2026-03-12 13:31:40.739714159 +0000 UTC m=+1319.188409555" lastFinishedPulling="2026-03-12 13:32:13.811907012 +0000 UTC m=+1352.260602408" observedRunningTime="2026-03-12 13:32:14.861220327 +0000 UTC m=+1353.309915753" watchObservedRunningTime="2026-03-12 13:32:14.863602524 +0000 UTC m=+1353.312297950" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.306175 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.371199 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8rkss"] Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.371446 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-8rkss" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerName="dnsmasq-dns" containerID="cri-o://cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb" gracePeriod=10 Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.809581 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.840807 4778 generic.go:334] "Generic (PLEG): container finished" podID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerID="cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb" exitCode=0 Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.841606 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8rkss" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.842020 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8rkss" event={"ID":"5dd405d8-c82b-49d0-a871-1c7c847638df","Type":"ContainerDied","Data":"cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb"} Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.842043 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8rkss" event={"ID":"5dd405d8-c82b-49d0-a871-1c7c847638df","Type":"ContainerDied","Data":"0cf55f4c77e0e83cbfd4fa4c9d04d1940beb400c64b78ffec689c21b7bd18ebf"} Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.842057 4778 scope.go:117] "RemoveContainer" containerID="cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.873345 4778 scope.go:117] "RemoveContainer" containerID="94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.896768 4778 scope.go:117] "RemoveContainer" containerID="cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb" Mar 12 13:32:15 crc kubenswrapper[4778]: E0312 13:32:15.897433 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb\": container with ID starting with cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb not found: ID does not exist" containerID="cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.897486 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb"} err="failed to get container status \"cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb\": rpc error: code = NotFound desc = could not find container \"cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb\": container with ID starting with cea929c8344637e6d5422f933285cd8d16eba93f79935ed8a5b3e6067be52dcb not found: ID does not exist" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.897512 4778 scope.go:117] "RemoveContainer" containerID="94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96" Mar 12 13:32:15 crc kubenswrapper[4778]: E0312 13:32:15.897790 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96\": container with ID starting with 94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96 not found: ID does not exist" containerID="94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.897835 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96"} err="failed to get container status \"94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96\": rpc error: code = NotFound desc = could not find container \"94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96\": container with ID starting with 94ff3282c8f419818bdb2d0b93c2c285da1b67c9dafa1b0134a3349197ba9d96 not found: ID does not exist" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.907424 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-config\") pod \"5dd405d8-c82b-49d0-a871-1c7c847638df\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.907465 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-dns-svc\") pod \"5dd405d8-c82b-49d0-a871-1c7c847638df\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.907583 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-nb\") pod \"5dd405d8-c82b-49d0-a871-1c7c847638df\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.907852 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-sb\") pod \"5dd405d8-c82b-49d0-a871-1c7c847638df\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.907918 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r26t5\" (UniqueName: \"kubernetes.io/projected/5dd405d8-c82b-49d0-a871-1c7c847638df-kube-api-access-r26t5\") pod \"5dd405d8-c82b-49d0-a871-1c7c847638df\" (UID: \"5dd405d8-c82b-49d0-a871-1c7c847638df\") " Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.913614 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dd405d8-c82b-49d0-a871-1c7c847638df-kube-api-access-r26t5" (OuterVolumeSpecName: "kube-api-access-r26t5") pod "5dd405d8-c82b-49d0-a871-1c7c847638df" (UID: "5dd405d8-c82b-49d0-a871-1c7c847638df"). InnerVolumeSpecName "kube-api-access-r26t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.954896 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-config" (OuterVolumeSpecName: "config") pod "5dd405d8-c82b-49d0-a871-1c7c847638df" (UID: "5dd405d8-c82b-49d0-a871-1c7c847638df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.956970 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5dd405d8-c82b-49d0-a871-1c7c847638df" (UID: "5dd405d8-c82b-49d0-a871-1c7c847638df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.960096 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5dd405d8-c82b-49d0-a871-1c7c847638df" (UID: "5dd405d8-c82b-49d0-a871-1c7c847638df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:15 crc kubenswrapper[4778]: I0312 13:32:15.971783 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5dd405d8-c82b-49d0-a871-1c7c847638df" (UID: "5dd405d8-c82b-49d0-a871-1c7c847638df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.011016 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.011058 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r26t5\" (UniqueName: \"kubernetes.io/projected/5dd405d8-c82b-49d0-a871-1c7c847638df-kube-api-access-r26t5\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.011071 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.011080 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.011088 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dd405d8-c82b-49d0-a871-1c7c847638df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.175205 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8rkss"] Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.182997 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8rkss"] Mar 12 13:32:16 crc kubenswrapper[4778]: I0312 13:32:16.267493 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" path="/var/lib/kubelet/pods/5dd405d8-c82b-49d0-a871-1c7c847638df/volumes" Mar 12 13:32:17 crc kubenswrapper[4778]: I0312 13:32:17.860945 4778 generic.go:334] "Generic (PLEG): container finished" podID="ec77eae6-4dac-4535-b0d3-98bd3422e4de" containerID="fc1fdc3b0586065e85920687a0b5a3f3a3005e79a719fda2a25493dca50c853e" exitCode=0 Mar 12 13:32:17 crc kubenswrapper[4778]: I0312 13:32:17.860992 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57cfm" event={"ID":"ec77eae6-4dac-4535-b0d3-98bd3422e4de","Type":"ContainerDied","Data":"fc1fdc3b0586065e85920687a0b5a3f3a3005e79a719fda2a25493dca50c853e"} Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.163197 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.261810 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-combined-ca-bundle\") pod \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.262236 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-config-data\") pod \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.262317 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlp2n\" (UniqueName: \"kubernetes.io/projected/ec77eae6-4dac-4535-b0d3-98bd3422e4de-kube-api-access-xlp2n\") pod \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\" (UID: \"ec77eae6-4dac-4535-b0d3-98bd3422e4de\") " Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.268006 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec77eae6-4dac-4535-b0d3-98bd3422e4de-kube-api-access-xlp2n" (OuterVolumeSpecName: "kube-api-access-xlp2n") pod "ec77eae6-4dac-4535-b0d3-98bd3422e4de" (UID: "ec77eae6-4dac-4535-b0d3-98bd3422e4de"). InnerVolumeSpecName "kube-api-access-xlp2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.286516 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec77eae6-4dac-4535-b0d3-98bd3422e4de" (UID: "ec77eae6-4dac-4535-b0d3-98bd3422e4de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.310728 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-config-data" (OuterVolumeSpecName: "config-data") pod "ec77eae6-4dac-4535-b0d3-98bd3422e4de" (UID: "ec77eae6-4dac-4535-b0d3-98bd3422e4de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.363686 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.363727 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec77eae6-4dac-4535-b0d3-98bd3422e4de-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.363738 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlp2n\" (UniqueName: \"kubernetes.io/projected/ec77eae6-4dac-4535-b0d3-98bd3422e4de-kube-api-access-xlp2n\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.880980 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-57cfm" event={"ID":"ec77eae6-4dac-4535-b0d3-98bd3422e4de","Type":"ContainerDied","Data":"e416653bf42b2d593f800644adfdb9dd57501c2cbfadea4a9a6bd3bc9f20d011"} Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.881024 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e416653bf42b2d593f800644adfdb9dd57501c2cbfadea4a9a6bd3bc9f20d011" Mar 12 13:32:19 crc kubenswrapper[4778]: I0312 13:32:19.881059 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-57cfm" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118114 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-f4n2l"] Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118706 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerName="init" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118718 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerName="init" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118733 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec77eae6-4dac-4535-b0d3-98bd3422e4de" containerName="keystone-db-sync" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118739 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec77eae6-4dac-4535-b0d3-98bd3422e4de" containerName="keystone-db-sync" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118747 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9793dfb5-c2a5-4dc1-993d-9e024a810ce8" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118753 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9793dfb5-c2a5-4dc1-993d-9e024a810ce8" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118764 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b694c81-3b07-45a1-9ca1-1e47e7430f1f" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118769 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b694c81-3b07-45a1-9ca1-1e47e7430f1f" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118783 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="729468a8-fded-4564-96c8-471d3cf48825" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118791 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="729468a8-fded-4564-96c8-471d3cf48825" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118806 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ff3988-1976-4049-8277-0acb36da44c5" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118811 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ff3988-1976-4049-8277-0acb36da44c5" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118822 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3800be73-3a09-42b6-8d01-592ccbc6aaa3" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118828 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3800be73-3a09-42b6-8d01-592ccbc6aaa3" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118841 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerName="dnsmasq-dns" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118846 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerName="dnsmasq-dns" Mar 12 13:32:20 crc kubenswrapper[4778]: E0312 13:32:20.118859 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f8bb53-a8a8-448f-8f42-349232e383ec" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.118866 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f8bb53-a8a8-448f-8f42-349232e383ec" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119029 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b694c81-3b07-45a1-9ca1-1e47e7430f1f" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119041 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f8bb53-a8a8-448f-8f42-349232e383ec" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119052 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ff3988-1976-4049-8277-0acb36da44c5" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119059 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dd405d8-c82b-49d0-a871-1c7c847638df" containerName="dnsmasq-dns" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119069 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9793dfb5-c2a5-4dc1-993d-9e024a810ce8" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119081 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3800be73-3a09-42b6-8d01-592ccbc6aaa3" containerName="mariadb-database-create" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119089 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="729468a8-fded-4564-96c8-471d3cf48825" containerName="mariadb-account-create-update" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.119098 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec77eae6-4dac-4535-b0d3-98bd3422e4de" containerName="keystone-db-sync" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.120251 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.142933 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-f4n2l"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.191686 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vhhp2"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.192973 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.195768 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.195957 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.196073 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rjpsk" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.196273 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.200484 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.222482 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vhhp2"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276697 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljj6s\" (UniqueName: \"kubernetes.io/projected/57227510-d79a-4924-941f-fdc35bda5d41-kube-api-access-ljj6s\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276742 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276777 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276803 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-scripts\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276837 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-credential-keys\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276885 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276911 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-combined-ca-bundle\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276944 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-svc\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276971 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-config\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.276996 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-fernet-keys\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.277019 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-config-data\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.277051 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxhsf\" (UniqueName: \"kubernetes.io/projected/5b157abf-4269-4449-8522-ac31cfbafd7e-kube-api-access-dxhsf\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.370277 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-d5pl9"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.371503 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.377842 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.377971 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5pxn8" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.378068 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379503 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljj6s\" (UniqueName: \"kubernetes.io/projected/57227510-d79a-4924-941f-fdc35bda5d41-kube-api-access-ljj6s\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379539 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379564 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379586 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-scripts\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379610 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-credential-keys\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379659 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379689 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-combined-ca-bundle\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379716 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-svc\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379738 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-config\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379759 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-fernet-keys\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379777 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-config-data\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.379814 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxhsf\" (UniqueName: \"kubernetes.io/projected/5b157abf-4269-4449-8522-ac31cfbafd7e-kube-api-access-dxhsf\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.380964 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.381343 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.381994 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.386119 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-svc\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.387457 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-config\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.393024 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-combined-ca-bundle\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.405444 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-scripts\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.409504 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-fernet-keys\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.410011 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-config-data\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.410052 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-d5pl9"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.426732 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-credential-keys\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.468649 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.482779 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxhsf\" (UniqueName: \"kubernetes.io/projected/5b157abf-4269-4449-8522-ac31cfbafd7e-kube-api-access-dxhsf\") pod \"dnsmasq-dns-5959f8865f-f4n2l\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.499320 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-config-data\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.499688 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-combined-ca-bundle\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.499729 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-db-sync-config-data\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.499816 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-scripts\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.499897 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpfhh\" (UniqueName: \"kubernetes.io/projected/bb110a1e-6281-437d-b857-eb79c4953e1a-kube-api-access-jpfhh\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.499958 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb110a1e-6281-437d-b857-eb79c4953e1a-etc-machine-id\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.526901 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljj6s\" (UniqueName: \"kubernetes.io/projected/57227510-d79a-4924-941f-fdc35bda5d41-kube-api-access-ljj6s\") pod \"keystone-bootstrap-vhhp2\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.533753 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.554692 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.589306 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.597610 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.622142 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpfhh\" (UniqueName: \"kubernetes.io/projected/bb110a1e-6281-437d-b857-eb79c4953e1a-kube-api-access-jpfhh\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.622232 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb110a1e-6281-437d-b857-eb79c4953e1a-etc-machine-id\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.622305 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-config-data\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.622343 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-combined-ca-bundle\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.622363 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-db-sync-config-data\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.622397 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-scripts\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.628827 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-combined-ca-bundle\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.629902 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb110a1e-6281-437d-b857-eb79c4953e1a-etc-machine-id\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.630840 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-db-sync-config-data\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.639203 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-scripts\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.644572 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-config-data\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.647348 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-6cvgs"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.648478 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.655732 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.656448 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.661462 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-d7pv5" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.669593 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.682343 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6cvgs"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.687167 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpfhh\" (UniqueName: \"kubernetes.io/projected/bb110a1e-6281-437d-b857-eb79c4953e1a-kube-api-access-jpfhh\") pod \"cinder-db-sync-d5pl9\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.692081 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-f4n2l"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.692885 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.700100 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c75fp"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.703397 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.714080 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-p59s9"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.721264 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.723712 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c75fp"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.725481 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-85xbx" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.725890 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.728974 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.729069 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-log-httpd\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.729111 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62wfc\" (UniqueName: \"kubernetes.io/projected/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-kube-api-access-62wfc\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.729143 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-config-data\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.729260 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-scripts\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.729350 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.729420 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-run-httpd\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.752202 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p59s9"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.769471 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zr86r"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.770997 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.774005 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.774070 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-72bvj" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.774196 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.780912 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zr86r"] Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831048 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-combined-ca-bundle\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831102 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831145 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz588\" (UniqueName: \"kubernetes.io/projected/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-kube-api-access-sz588\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831170 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-scripts\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831225 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-combined-ca-bundle\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831250 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831266 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-db-sync-config-data\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831286 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831317 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-run-httpd\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831335 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-config\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831357 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831373 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831401 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-config\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831424 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-log-httpd\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831443 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdgk6\" (UniqueName: \"kubernetes.io/projected/76f8f940-670d-47a0-a90a-afd3aa37a726-kube-api-access-qdgk6\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831463 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62wfc\" (UniqueName: \"kubernetes.io/projected/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-kube-api-access-62wfc\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831481 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5hmn\" (UniqueName: \"kubernetes.io/projected/a682334f-73c0-4e38-8f95-e5de661319bb-kube-api-access-r5hmn\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831500 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-config-data\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.831519 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.835635 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.837934 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-run-httpd\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.838231 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-log-httpd\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.842039 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.842797 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-config-data\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.850286 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.854020 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-scripts\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.862119 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62wfc\" (UniqueName: \"kubernetes.io/projected/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-kube-api-access-62wfc\") pod \"ceilometer-0\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935289 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-config\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935329 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935359 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-config\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935389 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdgk6\" (UniqueName: \"kubernetes.io/projected/76f8f940-670d-47a0-a90a-afd3aa37a726-kube-api-access-qdgk6\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935419 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5hmn\" (UniqueName: \"kubernetes.io/projected/a682334f-73c0-4e38-8f95-e5de661319bb-kube-api-access-r5hmn\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935440 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935459 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-combined-ca-bundle\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935488 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txchs\" (UniqueName: \"kubernetes.io/projected/faeb9cb3-46ae-428f-8c0e-538a2e552072-kube-api-access-txchs\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935525 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935540 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-config-data\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935567 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-combined-ca-bundle\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935592 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz588\" (UniqueName: \"kubernetes.io/projected/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-kube-api-access-sz588\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935620 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faeb9cb3-46ae-428f-8c0e-538a2e552072-logs\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935638 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-scripts\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935661 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-combined-ca-bundle\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935684 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-db-sync-config-data\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.935704 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.936650 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.937246 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-config\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.937731 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.940786 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-config\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.941546 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.941558 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.946219 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-db-sync-config-data\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.954126 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-combined-ca-bundle\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.957564 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz588\" (UniqueName: \"kubernetes.io/projected/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-kube-api-access-sz588\") pod \"dnsmasq-dns-58dd9ff6bc-c75fp\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.960425 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdgk6\" (UniqueName: \"kubernetes.io/projected/76f8f940-670d-47a0-a90a-afd3aa37a726-kube-api-access-qdgk6\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.960765 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5hmn\" (UniqueName: \"kubernetes.io/projected/a682334f-73c0-4e38-8f95-e5de661319bb-kube-api-access-r5hmn\") pod \"barbican-db-sync-p59s9\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.961227 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-combined-ca-bundle\") pod \"neutron-db-sync-6cvgs\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.980473 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:32:20 crc kubenswrapper[4778]: I0312 13:32:20.989970 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.037697 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txchs\" (UniqueName: \"kubernetes.io/projected/faeb9cb3-46ae-428f-8c0e-538a2e552072-kube-api-access-txchs\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.037740 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-config-data\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.037772 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-combined-ca-bundle\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.037808 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faeb9cb3-46ae-428f-8c0e-538a2e552072-logs\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.037825 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-scripts\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.039256 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faeb9cb3-46ae-428f-8c0e-538a2e552072-logs\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.042415 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-combined-ca-bundle\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.042458 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-scripts\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.047268 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-config-data\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.055224 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txchs\" (UniqueName: \"kubernetes.io/projected/faeb9cb3-46ae-428f-8c0e-538a2e552072-kube-api-access-txchs\") pod \"placement-db-sync-zr86r\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.064318 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.075563 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p59s9" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.088366 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.132614 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vhhp2"] Mar 12 13:32:21 crc kubenswrapper[4778]: W0312 13:32:21.187544 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57227510_d79a_4924_941f_fdc35bda5d41.slice/crio-b47ac6e700f6b26e79b8d33a0cd1c034fedfe7bd0e82190cbdb182f9948a680c WatchSource:0}: Error finding container b47ac6e700f6b26e79b8d33a0cd1c034fedfe7bd0e82190cbdb182f9948a680c: Status 404 returned error can't find the container with id b47ac6e700f6b26e79b8d33a0cd1c034fedfe7bd0e82190cbdb182f9948a680c Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.231799 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-f4n2l"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.562230 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-d5pl9"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.598450 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6cvgs"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.647400 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.871765 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c75fp"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.902757 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zr86r"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.908092 4778 generic.go:334] "Generic (PLEG): container finished" podID="5b157abf-4269-4449-8522-ac31cfbafd7e" containerID="57d636adb9d7ba52ab49822ca13b194593b944e05c770194c95b1d65f93a9998" exitCode=0 Mar 12 13:32:21 crc kubenswrapper[4778]: W0312 13:32:21.908297 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaeb9cb3_46ae_428f_8c0e_538a2e552072.slice/crio-d453594d6992bec0b731b36d1124f474724ec877404950823baad33e6f3bbe34 WatchSource:0}: Error finding container d453594d6992bec0b731b36d1124f474724ec877404950823baad33e6f3bbe34: Status 404 returned error can't find the container with id d453594d6992bec0b731b36d1124f474724ec877404950823baad33e6f3bbe34 Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.908323 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" event={"ID":"5b157abf-4269-4449-8522-ac31cfbafd7e","Type":"ContainerDied","Data":"57d636adb9d7ba52ab49822ca13b194593b944e05c770194c95b1d65f93a9998"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.908356 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" event={"ID":"5b157abf-4269-4449-8522-ac31cfbafd7e","Type":"ContainerStarted","Data":"2c8d5e8aae459cdcec524b4a37114dc4cbca1d6f2aa00764bfd89e2df611f32c"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.909504 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d5pl9" event={"ID":"bb110a1e-6281-437d-b857-eb79c4953e1a","Type":"ContainerStarted","Data":"8d37cd44357eb35c5c4917c8593f7e9902991ee071e5d92e025804bd35c2f76e"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.919671 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p59s9"] Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.921536 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhhp2" event={"ID":"57227510-d79a-4924-941f-fdc35bda5d41","Type":"ContainerStarted","Data":"59b401343563918013d35a2531aae9f420a7a4077e0d31999372fd3e7e21e169"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.921591 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhhp2" event={"ID":"57227510-d79a-4924-941f-fdc35bda5d41","Type":"ContainerStarted","Data":"b47ac6e700f6b26e79b8d33a0cd1c034fedfe7bd0e82190cbdb182f9948a680c"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.923915 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6cvgs" event={"ID":"76f8f940-670d-47a0-a90a-afd3aa37a726","Type":"ContainerStarted","Data":"856cfa1709bfc70905fa0560b8bcd9ee96d30c9ac3ff33d52f1608bcf34cd2fc"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.931438 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerStarted","Data":"a7e5a5f0fc47985a7306f104e3261cd746e20017382e7ac550b97742b3f6f6e4"} Mar 12 13:32:21 crc kubenswrapper[4778]: I0312 13:32:21.987866 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vhhp2" podStartSLOduration=1.987843971 podStartE2EDuration="1.987843971s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:21.964950892 +0000 UTC m=+1360.413646308" watchObservedRunningTime="2026-03-12 13:32:21.987843971 +0000 UTC m=+1360.436539367" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.243858 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.303757 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.371574 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-swift-storage-0\") pod \"5b157abf-4269-4449-8522-ac31cfbafd7e\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.371654 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxhsf\" (UniqueName: \"kubernetes.io/projected/5b157abf-4269-4449-8522-ac31cfbafd7e-kube-api-access-dxhsf\") pod \"5b157abf-4269-4449-8522-ac31cfbafd7e\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.371701 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-sb\") pod \"5b157abf-4269-4449-8522-ac31cfbafd7e\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.371728 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-nb\") pod \"5b157abf-4269-4449-8522-ac31cfbafd7e\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.371841 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-config\") pod \"5b157abf-4269-4449-8522-ac31cfbafd7e\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.371879 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-svc\") pod \"5b157abf-4269-4449-8522-ac31cfbafd7e\" (UID: \"5b157abf-4269-4449-8522-ac31cfbafd7e\") " Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.397454 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b157abf-4269-4449-8522-ac31cfbafd7e-kube-api-access-dxhsf" (OuterVolumeSpecName: "kube-api-access-dxhsf") pod "5b157abf-4269-4449-8522-ac31cfbafd7e" (UID: "5b157abf-4269-4449-8522-ac31cfbafd7e"). InnerVolumeSpecName "kube-api-access-dxhsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.399060 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b157abf-4269-4449-8522-ac31cfbafd7e" (UID: "5b157abf-4269-4449-8522-ac31cfbafd7e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.421867 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5b157abf-4269-4449-8522-ac31cfbafd7e" (UID: "5b157abf-4269-4449-8522-ac31cfbafd7e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.422206 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b157abf-4269-4449-8522-ac31cfbafd7e" (UID: "5b157abf-4269-4449-8522-ac31cfbafd7e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.454018 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b157abf-4269-4449-8522-ac31cfbafd7e" (UID: "5b157abf-4269-4449-8522-ac31cfbafd7e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.456818 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-config" (OuterVolumeSpecName: "config") pod "5b157abf-4269-4449-8522-ac31cfbafd7e" (UID: "5b157abf-4269-4449-8522-ac31cfbafd7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.474424 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.474464 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.474475 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.474488 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxhsf\" (UniqueName: \"kubernetes.io/projected/5b157abf-4269-4449-8522-ac31cfbafd7e-kube-api-access-dxhsf\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.474497 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.474505 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b157abf-4269-4449-8522-ac31cfbafd7e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.945877 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zr86r" event={"ID":"faeb9cb3-46ae-428f-8c0e-538a2e552072","Type":"ContainerStarted","Data":"d453594d6992bec0b731b36d1124f474724ec877404950823baad33e6f3bbe34"} Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.948944 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" event={"ID":"5b157abf-4269-4449-8522-ac31cfbafd7e","Type":"ContainerDied","Data":"2c8d5e8aae459cdcec524b4a37114dc4cbca1d6f2aa00764bfd89e2df611f32c"} Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.948986 4778 scope.go:117] "RemoveContainer" containerID="57d636adb9d7ba52ab49822ca13b194593b944e05c770194c95b1d65f93a9998" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.949121 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-f4n2l" Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.953945 4778 generic.go:334] "Generic (PLEG): container finished" podID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerID="deae0dcafd4096182fbb59df47a0f37084a6a4dd40f9ceb191771d7cc1d9e536" exitCode=0 Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.954008 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" event={"ID":"ed693bb0-f387-42e2-ae31-9ce01aee1cf9","Type":"ContainerDied","Data":"deae0dcafd4096182fbb59df47a0f37084a6a4dd40f9ceb191771d7cc1d9e536"} Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.954029 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" event={"ID":"ed693bb0-f387-42e2-ae31-9ce01aee1cf9","Type":"ContainerStarted","Data":"b878edd8e2f4ab4fa1fd5db083761fef998e508c23a74d8e3cd95838ec67e23c"} Mar 12 13:32:22 crc kubenswrapper[4778]: I0312 13:32:22.975786 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6cvgs" event={"ID":"76f8f940-670d-47a0-a90a-afd3aa37a726","Type":"ContainerStarted","Data":"86b41f2ea1c3794ed3e1fc975ecb18420f64bbd7611743de1aa319532e575758"} Mar 12 13:32:23 crc kubenswrapper[4778]: I0312 13:32:23.005929 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-6cvgs" podStartSLOduration=3.00591102 podStartE2EDuration="3.00591102s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:23.002585945 +0000 UTC m=+1361.451281341" watchObservedRunningTime="2026-03-12 13:32:23.00591102 +0000 UTC m=+1361.454606416" Mar 12 13:32:23 crc kubenswrapper[4778]: I0312 13:32:23.010432 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p59s9" event={"ID":"a682334f-73c0-4e38-8f95-e5de661319bb","Type":"ContainerStarted","Data":"9752a8239a23597303e4c0af125d25d5be143749ecb830c3912a0cbc8277763f"} Mar 12 13:32:23 crc kubenswrapper[4778]: I0312 13:32:23.148397 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-f4n2l"] Mar 12 13:32:23 crc kubenswrapper[4778]: I0312 13:32:23.157422 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-f4n2l"] Mar 12 13:32:24 crc kubenswrapper[4778]: I0312 13:32:24.024529 4778 generic.go:334] "Generic (PLEG): container finished" podID="befeb973-a1de-48f9-8de0-5559f75472dc" containerID="58438369e99b6009fb9ed545548de66fcc857634b3821d960d6e5735646c9d5c" exitCode=0 Mar 12 13:32:24 crc kubenswrapper[4778]: I0312 13:32:24.024877 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xg6z4" event={"ID":"befeb973-a1de-48f9-8de0-5559f75472dc","Type":"ContainerDied","Data":"58438369e99b6009fb9ed545548de66fcc857634b3821d960d6e5735646c9d5c"} Mar 12 13:32:24 crc kubenswrapper[4778]: I0312 13:32:24.034634 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" event={"ID":"ed693bb0-f387-42e2-ae31-9ce01aee1cf9","Type":"ContainerStarted","Data":"a86f0b8f75025d6f637f5995ea2db5120ec912d396c9c98099631c4e389118ac"} Mar 12 13:32:24 crc kubenswrapper[4778]: I0312 13:32:24.089094 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" podStartSLOduration=4.089071375 podStartE2EDuration="4.089071375s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:24.067873873 +0000 UTC m=+1362.516569279" watchObservedRunningTime="2026-03-12 13:32:24.089071375 +0000 UTC m=+1362.537766771" Mar 12 13:32:24 crc kubenswrapper[4778]: I0312 13:32:24.275087 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b157abf-4269-4449-8522-ac31cfbafd7e" path="/var/lib/kubelet/pods/5b157abf-4269-4449-8522-ac31cfbafd7e/volumes" Mar 12 13:32:25 crc kubenswrapper[4778]: I0312 13:32:25.042529 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:26 crc kubenswrapper[4778]: I0312 13:32:26.051531 4778 generic.go:334] "Generic (PLEG): container finished" podID="57227510-d79a-4924-941f-fdc35bda5d41" containerID="59b401343563918013d35a2531aae9f420a7a4077e0d31999372fd3e7e21e169" exitCode=0 Mar 12 13:32:26 crc kubenswrapper[4778]: I0312 13:32:26.051623 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhhp2" event={"ID":"57227510-d79a-4924-941f-fdc35bda5d41","Type":"ContainerDied","Data":"59b401343563918013d35a2531aae9f420a7a4077e0d31999372fd3e7e21e169"} Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.701976 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xg6z4" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.833707 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crgvn\" (UniqueName: \"kubernetes.io/projected/befeb973-a1de-48f9-8de0-5559f75472dc-kube-api-access-crgvn\") pod \"befeb973-a1de-48f9-8de0-5559f75472dc\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.833809 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-config-data\") pod \"befeb973-a1de-48f9-8de0-5559f75472dc\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.833950 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-db-sync-config-data\") pod \"befeb973-a1de-48f9-8de0-5559f75472dc\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.834052 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-combined-ca-bundle\") pod \"befeb973-a1de-48f9-8de0-5559f75472dc\" (UID: \"befeb973-a1de-48f9-8de0-5559f75472dc\") " Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.840260 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "befeb973-a1de-48f9-8de0-5559f75472dc" (UID: "befeb973-a1de-48f9-8de0-5559f75472dc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.847433 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/befeb973-a1de-48f9-8de0-5559f75472dc-kube-api-access-crgvn" (OuterVolumeSpecName: "kube-api-access-crgvn") pod "befeb973-a1de-48f9-8de0-5559f75472dc" (UID: "befeb973-a1de-48f9-8de0-5559f75472dc"). InnerVolumeSpecName "kube-api-access-crgvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.859691 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "befeb973-a1de-48f9-8de0-5559f75472dc" (UID: "befeb973-a1de-48f9-8de0-5559f75472dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.880843 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-config-data" (OuterVolumeSpecName: "config-data") pod "befeb973-a1de-48f9-8de0-5559f75472dc" (UID: "befeb973-a1de-48f9-8de0-5559f75472dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.936351 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.936383 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crgvn\" (UniqueName: \"kubernetes.io/projected/befeb973-a1de-48f9-8de0-5559f75472dc-kube-api-access-crgvn\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.936397 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:27 crc kubenswrapper[4778]: I0312 13:32:27.936406 4778 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/befeb973-a1de-48f9-8de0-5559f75472dc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:28 crc kubenswrapper[4778]: I0312 13:32:28.067027 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xg6z4" event={"ID":"befeb973-a1de-48f9-8de0-5559f75472dc","Type":"ContainerDied","Data":"f4635ea2bc5d2d0cce58645ef33f0143795167ef564ca2829fbc3740cec61b52"} Mar 12 13:32:28 crc kubenswrapper[4778]: I0312 13:32:28.067073 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4635ea2bc5d2d0cce58645ef33f0143795167ef564ca2829fbc3740cec61b52" Mar 12 13:32:28 crc kubenswrapper[4778]: I0312 13:32:28.067142 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xg6z4" Mar 12 13:32:28 crc kubenswrapper[4778]: I0312 13:32:28.557234 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:32:28 crc kubenswrapper[4778]: I0312 13:32:28.557592 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.676597 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c75fp"] Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.676897 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" containerID="cri-o://a86f0b8f75025d6f637f5995ea2db5120ec912d396c9c98099631c4e389118ac" gracePeriod=10 Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.684769 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.703481 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-v2vtk"] Mar 12 13:32:29 crc kubenswrapper[4778]: E0312 13:32:29.703840 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="befeb973-a1de-48f9-8de0-5559f75472dc" containerName="glance-db-sync" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.703853 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="befeb973-a1de-48f9-8de0-5559f75472dc" containerName="glance-db-sync" Mar 12 13:32:29 crc kubenswrapper[4778]: E0312 13:32:29.703867 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b157abf-4269-4449-8522-ac31cfbafd7e" containerName="init" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.703874 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b157abf-4269-4449-8522-ac31cfbafd7e" containerName="init" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.704028 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b157abf-4269-4449-8522-ac31cfbafd7e" containerName="init" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.704041 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="befeb973-a1de-48f9-8de0-5559f75472dc" containerName="glance-db-sync" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.704866 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.721139 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-v2vtk"] Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.744991 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.745068 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.745157 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.745256 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.745278 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7vfh\" (UniqueName: \"kubernetes.io/projected/39bd75fd-958e-4b3b-abd5-860adf376fd7-kube-api-access-p7vfh\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.745326 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-config\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.847025 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.847110 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.847136 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7vfh\" (UniqueName: \"kubernetes.io/projected/39bd75fd-958e-4b3b-abd5-860adf376fd7-kube-api-access-p7vfh\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.847184 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-config\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.847264 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.847308 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.848185 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.848691 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.849266 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.850075 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-config\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.854827 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:29 crc kubenswrapper[4778]: I0312 13:32:29.882467 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7vfh\" (UniqueName: \"kubernetes.io/projected/39bd75fd-958e-4b3b-abd5-860adf376fd7-kube-api-access-p7vfh\") pod \"dnsmasq-dns-785d8bcb8c-v2vtk\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.025594 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.582019 4778 generic.go:334] "Generic (PLEG): container finished" podID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerID="a86f0b8f75025d6f637f5995ea2db5120ec912d396c9c98099631c4e389118ac" exitCode=0 Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.582060 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" event={"ID":"ed693bb0-f387-42e2-ae31-9ce01aee1cf9","Type":"ContainerDied","Data":"a86f0b8f75025d6f637f5995ea2db5120ec912d396c9c98099631c4e389118ac"} Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.621881 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.623552 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.626034 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-l7l5j" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.626156 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.626463 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.631107 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.764892 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5xwn\" (UniqueName: \"kubernetes.io/projected/b533a505-eb7b-43a3-b95d-60cdc7198066-kube-api-access-f5xwn\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.765278 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.765304 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-scripts\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.765342 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-config-data\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.765394 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.765417 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-logs\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:30 crc kubenswrapper[4778]: I0312 13:32:30.765447 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.127533 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: connect: connection refused" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.134046 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.134117 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-scripts\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.136890 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-config-data\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.137035 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.137092 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-logs\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.137146 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.137226 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5xwn\" (UniqueName: \"kubernetes.io/projected/b533a505-eb7b-43a3-b95d-60cdc7198066-kube-api-access-f5xwn\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.139284 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-logs\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.139775 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.140656 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") device mount path \"/mnt/openstack/pv17\"" pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.142043 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-scripts\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.144092 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-config-data\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.156673 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.160654 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5xwn\" (UniqueName: \"kubernetes.io/projected/b533a505-eb7b-43a3-b95d-60cdc7198066-kube-api-access-f5xwn\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.162148 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.172767 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.179749 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.180071 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.211724 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.244413 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.340391 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.340464 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.340493 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.340514 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.340530 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.340575 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.341366 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbwgx\" (UniqueName: \"kubernetes.io/projected/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-kube-api-access-zbwgx\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443148 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443241 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbwgx\" (UniqueName: \"kubernetes.io/projected/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-kube-api-access-zbwgx\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443310 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443351 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443376 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443397 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443416 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.443511 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.444243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.444329 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.450385 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.450989 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.452503 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.467898 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbwgx\" (UniqueName: \"kubernetes.io/projected/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-kube-api-access-zbwgx\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.479933 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:31 crc kubenswrapper[4778]: I0312 13:32:31.591291 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:33 crc kubenswrapper[4778]: I0312 13:32:33.305132 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:33 crc kubenswrapper[4778]: I0312 13:32:33.424631 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:36 crc kubenswrapper[4778]: I0312 13:32:36.065310 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: connect: connection refused" Mar 12 13:32:37 crc kubenswrapper[4778]: E0312 13:32:37.251618 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Mar 12 13:32:37 crc kubenswrapper[4778]: E0312 13:32:37.252128 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5hmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-p59s9_openstack(a682334f-73c0-4e38-8f95-e5de661319bb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:32:37 crc kubenswrapper[4778]: E0312 13:32:37.253425 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-p59s9" podUID="a682334f-73c0-4e38-8f95-e5de661319bb" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.369419 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.470308 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-fernet-keys\") pod \"57227510-d79a-4924-941f-fdc35bda5d41\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.470668 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-scripts\") pod \"57227510-d79a-4924-941f-fdc35bda5d41\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.470707 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljj6s\" (UniqueName: \"kubernetes.io/projected/57227510-d79a-4924-941f-fdc35bda5d41-kube-api-access-ljj6s\") pod \"57227510-d79a-4924-941f-fdc35bda5d41\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.470795 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-combined-ca-bundle\") pod \"57227510-d79a-4924-941f-fdc35bda5d41\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.470858 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-config-data\") pod \"57227510-d79a-4924-941f-fdc35bda5d41\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.470899 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-credential-keys\") pod \"57227510-d79a-4924-941f-fdc35bda5d41\" (UID: \"57227510-d79a-4924-941f-fdc35bda5d41\") " Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.476439 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "57227510-d79a-4924-941f-fdc35bda5d41" (UID: "57227510-d79a-4924-941f-fdc35bda5d41"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.477454 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-scripts" (OuterVolumeSpecName: "scripts") pod "57227510-d79a-4924-941f-fdc35bda5d41" (UID: "57227510-d79a-4924-941f-fdc35bda5d41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.478616 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "57227510-d79a-4924-941f-fdc35bda5d41" (UID: "57227510-d79a-4924-941f-fdc35bda5d41"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.499053 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57227510-d79a-4924-941f-fdc35bda5d41-kube-api-access-ljj6s" (OuterVolumeSpecName: "kube-api-access-ljj6s") pod "57227510-d79a-4924-941f-fdc35bda5d41" (UID: "57227510-d79a-4924-941f-fdc35bda5d41"). InnerVolumeSpecName "kube-api-access-ljj6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.506977 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-config-data" (OuterVolumeSpecName: "config-data") pod "57227510-d79a-4924-941f-fdc35bda5d41" (UID: "57227510-d79a-4924-941f-fdc35bda5d41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.515336 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57227510-d79a-4924-941f-fdc35bda5d41" (UID: "57227510-d79a-4924-941f-fdc35bda5d41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.573160 4778 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.573214 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.573225 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljj6s\" (UniqueName: \"kubernetes.io/projected/57227510-d79a-4924-941f-fdc35bda5d41-kube-api-access-ljj6s\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.573236 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.573245 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.573253 4778 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57227510-d79a-4924-941f-fdc35bda5d41-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.641283 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhhp2" Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.641279 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhhp2" event={"ID":"57227510-d79a-4924-941f-fdc35bda5d41","Type":"ContainerDied","Data":"b47ac6e700f6b26e79b8d33a0cd1c034fedfe7bd0e82190cbdb182f9948a680c"} Mar 12 13:32:37 crc kubenswrapper[4778]: I0312 13:32:37.641329 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b47ac6e700f6b26e79b8d33a0cd1c034fedfe7bd0e82190cbdb182f9948a680c" Mar 12 13:32:37 crc kubenswrapper[4778]: E0312 13:32:37.646538 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-p59s9" podUID="a682334f-73c0-4e38-8f95-e5de661319bb" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.460461 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vhhp2"] Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.468414 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vhhp2"] Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.565017 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-56sfj"] Mar 12 13:32:38 crc kubenswrapper[4778]: E0312 13:32:38.565416 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57227510-d79a-4924-941f-fdc35bda5d41" containerName="keystone-bootstrap" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.565436 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="57227510-d79a-4924-941f-fdc35bda5d41" containerName="keystone-bootstrap" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.565636 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="57227510-d79a-4924-941f-fdc35bda5d41" containerName="keystone-bootstrap" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.566130 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.568000 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.568204 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.568540 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rjpsk" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.568696 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.572654 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.583033 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-56sfj"] Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.694958 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-credential-keys\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.695098 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-fernet-keys\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.695209 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94tqw\" (UniqueName: \"kubernetes.io/projected/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-kube-api-access-94tqw\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.695254 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-combined-ca-bundle\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.695311 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-config-data\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.695358 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-scripts\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.797021 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-fernet-keys\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.797102 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94tqw\" (UniqueName: \"kubernetes.io/projected/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-kube-api-access-94tqw\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.797137 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-combined-ca-bundle\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.797171 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-config-data\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.797259 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-scripts\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.797293 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-credential-keys\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.802756 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-credential-keys\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.802797 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-fernet-keys\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.803757 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-combined-ca-bundle\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.811829 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-config-data\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.812422 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-scripts\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.815902 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94tqw\" (UniqueName: \"kubernetes.io/projected/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-kube-api-access-94tqw\") pod \"keystone-bootstrap-56sfj\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:38 crc kubenswrapper[4778]: I0312 13:32:38.894880 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:40 crc kubenswrapper[4778]: I0312 13:32:40.276633 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57227510-d79a-4924-941f-fdc35bda5d41" path="/var/lib/kubelet/pods/57227510-d79a-4924-941f-fdc35bda5d41/volumes" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.120179 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.227976 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-nb\") pod \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.228072 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz588\" (UniqueName: \"kubernetes.io/projected/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-kube-api-access-sz588\") pod \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.228145 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-swift-storage-0\") pod \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.228241 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-config\") pod \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.228337 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-svc\") pod \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.228379 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-sb\") pod \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\" (UID: \"ed693bb0-f387-42e2-ae31-9ce01aee1cf9\") " Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.233924 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-kube-api-access-sz588" (OuterVolumeSpecName: "kube-api-access-sz588") pod "ed693bb0-f387-42e2-ae31-9ce01aee1cf9" (UID: "ed693bb0-f387-42e2-ae31-9ce01aee1cf9"). InnerVolumeSpecName "kube-api-access-sz588". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.280032 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-config" (OuterVolumeSpecName: "config") pod "ed693bb0-f387-42e2-ae31-9ce01aee1cf9" (UID: "ed693bb0-f387-42e2-ae31-9ce01aee1cf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.282889 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ed693bb0-f387-42e2-ae31-9ce01aee1cf9" (UID: "ed693bb0-f387-42e2-ae31-9ce01aee1cf9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.283034 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed693bb0-f387-42e2-ae31-9ce01aee1cf9" (UID: "ed693bb0-f387-42e2-ae31-9ce01aee1cf9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.290096 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed693bb0-f387-42e2-ae31-9ce01aee1cf9" (UID: "ed693bb0-f387-42e2-ae31-9ce01aee1cf9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.293908 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed693bb0-f387-42e2-ae31-9ce01aee1cf9" (UID: "ed693bb0-f387-42e2-ae31-9ce01aee1cf9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.331707 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.332194 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.332225 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.332324 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz588\" (UniqueName: \"kubernetes.io/projected/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-kube-api-access-sz588\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.332344 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.332354 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed693bb0-f387-42e2-ae31-9ce01aee1cf9-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.727207 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" event={"ID":"ed693bb0-f387-42e2-ae31-9ce01aee1cf9","Type":"ContainerDied","Data":"b878edd8e2f4ab4fa1fd5db083761fef998e508c23a74d8e3cd95838ec67e23c"} Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.727285 4778 scope.go:117] "RemoveContainer" containerID="a86f0b8f75025d6f637f5995ea2db5120ec912d396c9c98099631c4e389118ac" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.727519 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.775597 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c75fp"] Mar 12 13:32:45 crc kubenswrapper[4778]: I0312 13:32:45.785253 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c75fp"] Mar 12 13:32:46 crc kubenswrapper[4778]: I0312 13:32:46.066176 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-c75fp" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: i/o timeout" Mar 12 13:32:46 crc kubenswrapper[4778]: I0312 13:32:46.266851 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" path="/var/lib/kubelet/pods/ed693bb0-f387-42e2-ae31-9ce01aee1cf9/volumes" Mar 12 13:32:46 crc kubenswrapper[4778]: E0312 13:32:46.558343 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 12 13:32:46 crc kubenswrapper[4778]: E0312 13:32:46.558816 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jpfhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-d5pl9_openstack(bb110a1e-6281-437d-b857-eb79c4953e1a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 13:32:46 crc kubenswrapper[4778]: E0312 13:32:46.559957 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-d5pl9" podUID="bb110a1e-6281-437d-b857-eb79c4953e1a" Mar 12 13:32:46 crc kubenswrapper[4778]: I0312 13:32:46.584432 4778 scope.go:117] "RemoveContainer" containerID="deae0dcafd4096182fbb59df47a0f37084a6a4dd40f9ceb191771d7cc1d9e536" Mar 12 13:32:46 crc kubenswrapper[4778]: E0312 13:32:46.758337 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-d5pl9" podUID="bb110a1e-6281-437d-b857-eb79c4953e1a" Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.070401 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-56sfj"] Mar 12 13:32:47 crc kubenswrapper[4778]: W0312 13:32:47.087146 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1af573ef_51c3_4bfc_8de6_eb1be8b75c76.slice/crio-ebc59f76f06ba10050abce1212b94660f06fec69cccf436b21f8ae1838a2520b WatchSource:0}: Error finding container ebc59f76f06ba10050abce1212b94660f06fec69cccf436b21f8ae1838a2520b: Status 404 returned error can't find the container with id ebc59f76f06ba10050abce1212b94660f06fec69cccf436b21f8ae1838a2520b Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.092940 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.152279 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-v2vtk"] Mar 12 13:32:47 crc kubenswrapper[4778]: W0312 13:32:47.155963 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39bd75fd_958e_4b3b_abd5_860adf376fd7.slice/crio-415522e7cc2372bb11dfe09957497d4a3efac5b28086b59aebe2586918e3f99d WatchSource:0}: Error finding container 415522e7cc2372bb11dfe09957497d4a3efac5b28086b59aebe2586918e3f99d: Status 404 returned error can't find the container with id 415522e7cc2372bb11dfe09957497d4a3efac5b28086b59aebe2586918e3f99d Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.177661 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:47 crc kubenswrapper[4778]: W0312 13:32:47.179471 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa998ea4_f50d_4441_b6ad_b160a19ea4a9.slice/crio-3c8eaac29f690ea90c0a8b68198ea45e6be61ad552e67a0c3c11ec5342477745 WatchSource:0}: Error finding container 3c8eaac29f690ea90c0a8b68198ea45e6be61ad552e67a0c3c11ec5342477745: Status 404 returned error can't find the container with id 3c8eaac29f690ea90c0a8b68198ea45e6be61ad552e67a0c3c11ec5342477745 Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.298607 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.802612 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-56sfj" event={"ID":"1af573ef-51c3-4bfc-8de6-eb1be8b75c76","Type":"ContainerStarted","Data":"710035f2fd1c6ce07427dd61579057ea7d418eb1c9532e9c2ad2d414dc76cbb9"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.802845 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-56sfj" event={"ID":"1af573ef-51c3-4bfc-8de6-eb1be8b75c76","Type":"ContainerStarted","Data":"ebc59f76f06ba10050abce1212b94660f06fec69cccf436b21f8ae1838a2520b"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.805212 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa998ea4-f50d-4441-b6ad-b160a19ea4a9","Type":"ContainerStarted","Data":"3c8eaac29f690ea90c0a8b68198ea45e6be61ad552e67a0c3c11ec5342477745"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.820643 4778 generic.go:334] "Generic (PLEG): container finished" podID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerID="7c88372c4eebf35fa3a0e19eba355c02e9d34ad468328fc457e997e453d917f3" exitCode=0 Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.820730 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" event={"ID":"39bd75fd-958e-4b3b-abd5-860adf376fd7","Type":"ContainerDied","Data":"7c88372c4eebf35fa3a0e19eba355c02e9d34ad468328fc457e997e453d917f3"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.820757 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" event={"ID":"39bd75fd-958e-4b3b-abd5-860adf376fd7","Type":"ContainerStarted","Data":"415522e7cc2372bb11dfe09957497d4a3efac5b28086b59aebe2586918e3f99d"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.823615 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zr86r" event={"ID":"faeb9cb3-46ae-428f-8c0e-538a2e552072","Type":"ContainerStarted","Data":"434f9dbc426c8bc5145f54de2b34c16cd91006660bd978fe7ad9311fc8579e69"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.827946 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerStarted","Data":"a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.839631 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b533a505-eb7b-43a3-b95d-60cdc7198066","Type":"ContainerStarted","Data":"bc3d154f21afd30a55fef211dca4fe535b5cb4020d2ea7a7c87c369860d2b039"} Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.872842 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-56sfj" podStartSLOduration=9.87282081 podStartE2EDuration="9.87282081s" podCreationTimestamp="2026-03-12 13:32:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:47.838256159 +0000 UTC m=+1386.286951555" watchObservedRunningTime="2026-03-12 13:32:47.87282081 +0000 UTC m=+1386.321516206" Mar 12 13:32:47 crc kubenswrapper[4778]: I0312 13:32:47.888404 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zr86r" podStartSLOduration=3.276943037 podStartE2EDuration="27.888389831s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="2026-03-12 13:32:21.910840787 +0000 UTC m=+1360.359536183" lastFinishedPulling="2026-03-12 13:32:46.522287541 +0000 UTC m=+1384.970982977" observedRunningTime="2026-03-12 13:32:47.887942189 +0000 UTC m=+1386.336637585" watchObservedRunningTime="2026-03-12 13:32:47.888389831 +0000 UTC m=+1386.337085227" Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.852764 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa998ea4-f50d-4441-b6ad-b160a19ea4a9","Type":"ContainerStarted","Data":"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a"} Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.853356 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa998ea4-f50d-4441-b6ad-b160a19ea4a9","Type":"ContainerStarted","Data":"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513"} Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.853147 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-httpd" containerID="cri-o://9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a" gracePeriod=30 Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.852949 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-log" containerID="cri-o://34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513" gracePeriod=30 Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.858700 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" event={"ID":"39bd75fd-958e-4b3b-abd5-860adf376fd7","Type":"ContainerStarted","Data":"cc6fc61a82e88c3140b3629f45196f98ee08d5f2fdb0df9b40fe66806a0ccbfd"} Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.859997 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.862207 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b533a505-eb7b-43a3-b95d-60cdc7198066","Type":"ContainerStarted","Data":"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34"} Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.862243 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b533a505-eb7b-43a3-b95d-60cdc7198066","Type":"ContainerStarted","Data":"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d"} Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.862490 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-log" containerID="cri-o://d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d" gracePeriod=30 Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.862512 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-httpd" containerID="cri-o://d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34" gracePeriod=30 Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.892681 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=19.892654858 podStartE2EDuration="19.892654858s" podCreationTimestamp="2026-03-12 13:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:48.880462753 +0000 UTC m=+1387.329158179" watchObservedRunningTime="2026-03-12 13:32:48.892654858 +0000 UTC m=+1387.341350264" Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.924125 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=19.92410444 podStartE2EDuration="19.92410444s" podCreationTimestamp="2026-03-12 13:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:48.921893238 +0000 UTC m=+1387.370588644" watchObservedRunningTime="2026-03-12 13:32:48.92410444 +0000 UTC m=+1387.372799836" Mar 12 13:32:48 crc kubenswrapper[4778]: I0312 13:32:48.950516 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" podStartSLOduration=19.950494119 podStartE2EDuration="19.950494119s" podCreationTimestamp="2026-03-12 13:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:48.941801932 +0000 UTC m=+1387.390497328" watchObservedRunningTime="2026-03-12 13:32:48.950494119 +0000 UTC m=+1387.399189515" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.615884 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.722500 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.722668 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-httpd-run\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.722726 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5xwn\" (UniqueName: \"kubernetes.io/projected/b533a505-eb7b-43a3-b95d-60cdc7198066-kube-api-access-f5xwn\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723259 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-config-data\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723307 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-logs\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723304 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723333 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-scripts\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723373 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-combined-ca-bundle\") pod \"b533a505-eb7b-43a3-b95d-60cdc7198066\" (UID: \"b533a505-eb7b-43a3-b95d-60cdc7198066\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723502 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-logs" (OuterVolumeSpecName: "logs") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723919 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.723945 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b533a505-eb7b-43a3-b95d-60cdc7198066-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.732358 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.732362 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-scripts" (OuterVolumeSpecName: "scripts") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.734264 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b533a505-eb7b-43a3-b95d-60cdc7198066-kube-api-access-f5xwn" (OuterVolumeSpecName: "kube-api-access-f5xwn") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "kube-api-access-f5xwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.751532 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.789854 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-config-data" (OuterVolumeSpecName: "config-data") pod "b533a505-eb7b-43a3-b95d-60cdc7198066" (UID: "b533a505-eb7b-43a3-b95d-60cdc7198066"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.794451 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.825511 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5xwn\" (UniqueName: \"kubernetes.io/projected/b533a505-eb7b-43a3-b95d-60cdc7198066-kube-api-access-f5xwn\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.825546 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.825555 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.825565 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b533a505-eb7b-43a3-b95d-60cdc7198066-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.825590 4778 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.841537 4778 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.871374 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerStarted","Data":"f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873650 4778 generic.go:334] "Generic (PLEG): container finished" podID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerID="d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34" exitCode=143 Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873678 4778 generic.go:334] "Generic (PLEG): container finished" podID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerID="d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d" exitCode=143 Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873719 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b533a505-eb7b-43a3-b95d-60cdc7198066","Type":"ContainerDied","Data":"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873734 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873753 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b533a505-eb7b-43a3-b95d-60cdc7198066","Type":"ContainerDied","Data":"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873766 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b533a505-eb7b-43a3-b95d-60cdc7198066","Type":"ContainerDied","Data":"bc3d154f21afd30a55fef211dca4fe535b5cb4020d2ea7a7c87c369860d2b039"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.873772 4778 scope.go:117] "RemoveContainer" containerID="d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.875411 4778 generic.go:334] "Generic (PLEG): container finished" podID="76f8f940-670d-47a0-a90a-afd3aa37a726" containerID="86b41f2ea1c3794ed3e1fc975ecb18420f64bbd7611743de1aa319532e575758" exitCode=0 Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.875459 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6cvgs" event={"ID":"76f8f940-670d-47a0-a90a-afd3aa37a726","Type":"ContainerDied","Data":"86b41f2ea1c3794ed3e1fc975ecb18420f64bbd7611743de1aa319532e575758"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.879502 4778 generic.go:334] "Generic (PLEG): container finished" podID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerID="9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a" exitCode=0 Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.879652 4778 generic.go:334] "Generic (PLEG): container finished" podID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerID="34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513" exitCode=143 Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.879569 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa998ea4-f50d-4441-b6ad-b160a19ea4a9","Type":"ContainerDied","Data":"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.879625 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.879711 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa998ea4-f50d-4441-b6ad-b160a19ea4a9","Type":"ContainerDied","Data":"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.879923 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa998ea4-f50d-4441-b6ad-b160a19ea4a9","Type":"ContainerDied","Data":"3c8eaac29f690ea90c0a8b68198ea45e6be61ad552e67a0c3c11ec5342477745"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.881505 4778 generic.go:334] "Generic (PLEG): container finished" podID="faeb9cb3-46ae-428f-8c0e-538a2e552072" containerID="434f9dbc426c8bc5145f54de2b34c16cd91006660bd978fe7ad9311fc8579e69" exitCode=0 Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.882693 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zr86r" event={"ID":"faeb9cb3-46ae-428f-8c0e-538a2e552072","Type":"ContainerDied","Data":"434f9dbc426c8bc5145f54de2b34c16cd91006660bd978fe7ad9311fc8579e69"} Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.906944 4778 scope.go:117] "RemoveContainer" containerID="d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926150 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-config-data\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926256 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-httpd-run\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926302 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-logs\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926346 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbwgx\" (UniqueName: \"kubernetes.io/projected/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-kube-api-access-zbwgx\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926373 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-scripts\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926405 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.926453 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-combined-ca-bundle\") pod \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\" (UID: \"aa998ea4-f50d-4441-b6ad-b160a19ea4a9\") " Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.927486 4778 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.929047 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.929270 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-logs" (OuterVolumeSpecName: "logs") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.932507 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-kube-api-access-zbwgx" (OuterVolumeSpecName: "kube-api-access-zbwgx") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "kube-api-access-zbwgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.932518 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-scripts" (OuterVolumeSpecName: "scripts") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.933463 4778 scope.go:117] "RemoveContainer" containerID="d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34" Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.934007 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34\": container with ID starting with d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34 not found: ID does not exist" containerID="d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934042 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34"} err="failed to get container status \"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34\": rpc error: code = NotFound desc = could not find container \"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34\": container with ID starting with d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34 not found: ID does not exist" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934069 4778 scope.go:117] "RemoveContainer" containerID="d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934413 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.934485 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d\": container with ID starting with d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d not found: ID does not exist" containerID="d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934525 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d"} err="failed to get container status \"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d\": rpc error: code = NotFound desc = could not find container \"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d\": container with ID starting with d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d not found: ID does not exist" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934542 4778 scope.go:117] "RemoveContainer" containerID="d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934817 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34"} err="failed to get container status \"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34\": rpc error: code = NotFound desc = could not find container \"d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34\": container with ID starting with d0d48a7ad8664426390a968f02d1600a2fa102d2c72f9c248494c7e0624b1b34 not found: ID does not exist" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.934841 4778 scope.go:117] "RemoveContainer" containerID="d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.935118 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d"} err="failed to get container status \"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d\": rpc error: code = NotFound desc = could not find container \"d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d\": container with ID starting with d4d7c6b9984a293eb0854447833f3f9cb59361882500cfb456fd7b159e0fba3d not found: ID does not exist" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.935138 4778 scope.go:117] "RemoveContainer" containerID="9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.941744 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.949510 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.967248 4778 scope.go:117] "RemoveContainer" containerID="34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.968354 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.968806 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-httpd" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.968826 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-httpd" Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.969365 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="init" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969378 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="init" Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.969403 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-log" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969410 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-log" Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.969421 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-log" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969428 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-log" Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.969463 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969471 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" Mar 12 13:32:49 crc kubenswrapper[4778]: E0312 13:32:49.969478 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-httpd" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969484 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-httpd" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969799 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-log" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969829 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed693bb0-f387-42e2-ae31-9ce01aee1cf9" containerName="dnsmasq-dns" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969847 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-log" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969857 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" containerName="glance-httpd" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.969866 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" containerName="glance-httpd" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.970808 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.975885 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.979377 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.984351 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:49 crc kubenswrapper[4778]: I0312 13:32:49.986913 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.002650 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-config-data" (OuterVolumeSpecName: "config-data") pod "aa998ea4-f50d-4441-b6ad-b160a19ea4a9" (UID: "aa998ea4-f50d-4441-b6ad-b160a19ea4a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.016966 4778 scope.go:117] "RemoveContainer" containerID="9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a" Mar 12 13:32:50 crc kubenswrapper[4778]: E0312 13:32:50.017474 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a\": container with ID starting with 9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a not found: ID does not exist" containerID="9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.017553 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a"} err="failed to get container status \"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a\": rpc error: code = NotFound desc = could not find container \"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a\": container with ID starting with 9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a not found: ID does not exist" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.017586 4778 scope.go:117] "RemoveContainer" containerID="34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513" Mar 12 13:32:50 crc kubenswrapper[4778]: E0312 13:32:50.017900 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513\": container with ID starting with 34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513 not found: ID does not exist" containerID="34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.017924 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513"} err="failed to get container status \"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513\": rpc error: code = NotFound desc = could not find container \"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513\": container with ID starting with 34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513 not found: ID does not exist" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.017938 4778 scope.go:117] "RemoveContainer" containerID="9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.018278 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a"} err="failed to get container status \"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a\": rpc error: code = NotFound desc = could not find container \"9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a\": container with ID starting with 9f2e7f69bdd6233212da78ad84aba66a12c2daa431d1cfce5136f9137ccffb2a not found: ID does not exist" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.018296 4778 scope.go:117] "RemoveContainer" containerID="34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.018506 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513"} err="failed to get container status \"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513\": rpc error: code = NotFound desc = could not find container \"34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513\": container with ID starting with 34daa0a884944841a846bd99aab0d1c3b6985cc49de2820be6feb9d362176513 not found: ID does not exist" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029064 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029090 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029099 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029106 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029115 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbwgx\" (UniqueName: \"kubernetes.io/projected/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-kube-api-access-zbwgx\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029123 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa998ea4-f50d-4441-b6ad-b160a19ea4a9-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.029144 4778 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.052209 4778 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.130807 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.130887 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.130917 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.130937 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pklz\" (UniqueName: \"kubernetes.io/projected/ac92f5c5-e457-4915-a919-0dbe3df23ce8-kube-api-access-5pklz\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.130973 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.131028 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.131107 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.131204 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-logs\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.131321 4778 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.229493 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.234876 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.234939 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.234964 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.234985 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pklz\" (UniqueName: \"kubernetes.io/projected/ac92f5c5-e457-4915-a919-0dbe3df23ce8-kube-api-access-5pklz\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.235008 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.235031 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.235052 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.235099 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-logs\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.235655 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-logs\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.236424 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") device mount path \"/mnt/openstack/pv17\"" pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.237163 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.240986 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.249282 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.251923 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.252854 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.256690 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pklz\" (UniqueName: \"kubernetes.io/projected/ac92f5c5-e457-4915-a919-0dbe3df23ce8-kube-api-access-5pklz\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.290556 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b533a505-eb7b-43a3-b95d-60cdc7198066" path="/var/lib/kubelet/pods/b533a505-eb7b-43a3-b95d-60cdc7198066/volumes" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.304319 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.307496 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.311661 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.317533 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.320180 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.320409 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.326432 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.440906 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfwzz\" (UniqueName: \"kubernetes.io/projected/c2b13038-d271-48f5-bd28-a38e2b9dff02-kube-api-access-xfwzz\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441066 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441152 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441224 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441457 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441498 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441543 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.441596 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-logs\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543604 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfwzz\" (UniqueName: \"kubernetes.io/projected/c2b13038-d271-48f5-bd28-a38e2b9dff02-kube-api-access-xfwzz\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543676 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543695 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543715 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543759 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543782 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543805 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.543828 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-logs\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.544329 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-logs\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.545579 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.552860 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.554561 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.554943 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.557082 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.557801 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.561628 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfwzz\" (UniqueName: \"kubernetes.io/projected/c2b13038-d271-48f5-bd28-a38e2b9dff02-kube-api-access-xfwzz\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.575226 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.601618 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.640232 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.899707 4778 generic.go:334] "Generic (PLEG): container finished" podID="1af573ef-51c3-4bfc-8de6-eb1be8b75c76" containerID="710035f2fd1c6ce07427dd61579057ea7d418eb1c9532e9c2ad2d414dc76cbb9" exitCode=0 Mar 12 13:32:50 crc kubenswrapper[4778]: I0312 13:32:50.900118 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-56sfj" event={"ID":"1af573ef-51c3-4bfc-8de6-eb1be8b75c76","Type":"ContainerDied","Data":"710035f2fd1c6ce07427dd61579057ea7d418eb1c9532e9c2ad2d414dc76cbb9"} Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.130038 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:32:51 crc kubenswrapper[4778]: W0312 13:32:51.153247 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac92f5c5_e457_4915_a919_0dbe3df23ce8.slice/crio-de7bb235534c3c0c1a6530e35fd6d03d222f02129ca88b49fda3a8c136ab05b7 WatchSource:0}: Error finding container de7bb235534c3c0c1a6530e35fd6d03d222f02129ca88b49fda3a8c136ab05b7: Status 404 returned error can't find the container with id de7bb235534c3c0c1a6530e35fd6d03d222f02129ca88b49fda3a8c136ab05b7 Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.239983 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.277933 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.348745 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374264 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-combined-ca-bundle\") pod \"76f8f940-670d-47a0-a90a-afd3aa37a726\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374348 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-config\") pod \"76f8f940-670d-47a0-a90a-afd3aa37a726\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374445 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-combined-ca-bundle\") pod \"faeb9cb3-46ae-428f-8c0e-538a2e552072\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374500 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faeb9cb3-46ae-428f-8c0e-538a2e552072-logs\") pod \"faeb9cb3-46ae-428f-8c0e-538a2e552072\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374579 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-scripts\") pod \"faeb9cb3-46ae-428f-8c0e-538a2e552072\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374627 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdgk6\" (UniqueName: \"kubernetes.io/projected/76f8f940-670d-47a0-a90a-afd3aa37a726-kube-api-access-qdgk6\") pod \"76f8f940-670d-47a0-a90a-afd3aa37a726\" (UID: \"76f8f940-670d-47a0-a90a-afd3aa37a726\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374654 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txchs\" (UniqueName: \"kubernetes.io/projected/faeb9cb3-46ae-428f-8c0e-538a2e552072-kube-api-access-txchs\") pod \"faeb9cb3-46ae-428f-8c0e-538a2e552072\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.374731 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-config-data\") pod \"faeb9cb3-46ae-428f-8c0e-538a2e552072\" (UID: \"faeb9cb3-46ae-428f-8c0e-538a2e552072\") " Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.377242 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faeb9cb3-46ae-428f-8c0e-538a2e552072-logs" (OuterVolumeSpecName: "logs") pod "faeb9cb3-46ae-428f-8c0e-538a2e552072" (UID: "faeb9cb3-46ae-428f-8c0e-538a2e552072"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.381465 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f8f940-670d-47a0-a90a-afd3aa37a726-kube-api-access-qdgk6" (OuterVolumeSpecName: "kube-api-access-qdgk6") pod "76f8f940-670d-47a0-a90a-afd3aa37a726" (UID: "76f8f940-670d-47a0-a90a-afd3aa37a726"). InnerVolumeSpecName "kube-api-access-qdgk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.387483 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faeb9cb3-46ae-428f-8c0e-538a2e552072-kube-api-access-txchs" (OuterVolumeSpecName: "kube-api-access-txchs") pod "faeb9cb3-46ae-428f-8c0e-538a2e552072" (UID: "faeb9cb3-46ae-428f-8c0e-538a2e552072"). InnerVolumeSpecName "kube-api-access-txchs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.389541 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-scripts" (OuterVolumeSpecName: "scripts") pod "faeb9cb3-46ae-428f-8c0e-538a2e552072" (UID: "faeb9cb3-46ae-428f-8c0e-538a2e552072"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.404277 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "faeb9cb3-46ae-428f-8c0e-538a2e552072" (UID: "faeb9cb3-46ae-428f-8c0e-538a2e552072"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.405682 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-config" (OuterVolumeSpecName: "config") pod "76f8f940-670d-47a0-a90a-afd3aa37a726" (UID: "76f8f940-670d-47a0-a90a-afd3aa37a726"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.412929 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76f8f940-670d-47a0-a90a-afd3aa37a726" (UID: "76f8f940-670d-47a0-a90a-afd3aa37a726"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.416781 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-config-data" (OuterVolumeSpecName: "config-data") pod "faeb9cb3-46ae-428f-8c0e-538a2e552072" (UID: "faeb9cb3-46ae-428f-8c0e-538a2e552072"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476807 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476836 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/76f8f940-670d-47a0-a90a-afd3aa37a726-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476846 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476855 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faeb9cb3-46ae-428f-8c0e-538a2e552072-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476864 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476872 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdgk6\" (UniqueName: \"kubernetes.io/projected/76f8f940-670d-47a0-a90a-afd3aa37a726-kube-api-access-qdgk6\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476884 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txchs\" (UniqueName: \"kubernetes.io/projected/faeb9cb3-46ae-428f-8c0e-538a2e552072-kube-api-access-txchs\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.476893 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faeb9cb3-46ae-428f-8c0e-538a2e552072-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.950199 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6cvgs" event={"ID":"76f8f940-670d-47a0-a90a-afd3aa37a726","Type":"ContainerDied","Data":"856cfa1709bfc70905fa0560b8bcd9ee96d30c9ac3ff33d52f1608bcf34cd2fc"} Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.950603 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="856cfa1709bfc70905fa0560b8bcd9ee96d30c9ac3ff33d52f1608bcf34cd2fc" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.950237 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6cvgs" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.952295 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c2b13038-d271-48f5-bd28-a38e2b9dff02","Type":"ContainerStarted","Data":"15eaae81b5ec94e32bcb75db667617fbe51c32c5f0cac153a8a191ff89576b97"} Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.958948 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac92f5c5-e457-4915-a919-0dbe3df23ce8","Type":"ContainerStarted","Data":"0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260"} Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.958987 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac92f5c5-e457-4915-a919-0dbe3df23ce8","Type":"ContainerStarted","Data":"de7bb235534c3c0c1a6530e35fd6d03d222f02129ca88b49fda3a8c136ab05b7"} Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.960836 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zr86r" event={"ID":"faeb9cb3-46ae-428f-8c0e-538a2e552072","Type":"ContainerDied","Data":"d453594d6992bec0b731b36d1124f474724ec877404950823baad33e6f3bbe34"} Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.960868 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d453594d6992bec0b731b36d1124f474724ec877404950823baad33e6f3bbe34" Mar 12 13:32:51 crc kubenswrapper[4778]: I0312 13:32:51.960973 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zr86r" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.075660 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-79ccdbbbbd-gl27l"] Mar 12 13:32:52 crc kubenswrapper[4778]: E0312 13:32:52.077141 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f8f940-670d-47a0-a90a-afd3aa37a726" containerName="neutron-db-sync" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.077157 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f8f940-670d-47a0-a90a-afd3aa37a726" containerName="neutron-db-sync" Mar 12 13:32:52 crc kubenswrapper[4778]: E0312 13:32:52.077199 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faeb9cb3-46ae-428f-8c0e-538a2e552072" containerName="placement-db-sync" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.077206 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="faeb9cb3-46ae-428f-8c0e-538a2e552072" containerName="placement-db-sync" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.077354 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f8f940-670d-47a0-a90a-afd3aa37a726" containerName="neutron-db-sync" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.077365 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="faeb9cb3-46ae-428f-8c0e-538a2e552072" containerName="placement-db-sync" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.078143 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.082238 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.082472 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.083706 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.083901 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-72bvj" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.084048 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.155223 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-v2vtk"] Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.155543 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="dnsmasq-dns" containerID="cri-o://cc6fc61a82e88c3140b3629f45196f98ee08d5f2fdb0df9b40fe66806a0ccbfd" gracePeriod=10 Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197015 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-public-tls-certs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197118 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68092e68-04e5-4530-8d94-859789faeb94-logs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197142 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-config-data\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197162 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnrxw\" (UniqueName: \"kubernetes.io/projected/68092e68-04e5-4530-8d94-859789faeb94-kube-api-access-jnrxw\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197203 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-internal-tls-certs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197238 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-scripts\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.197515 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-combined-ca-bundle\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303009 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-public-tls-certs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303170 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68092e68-04e5-4530-8d94-859789faeb94-logs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303213 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-config-data\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303241 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnrxw\" (UniqueName: \"kubernetes.io/projected/68092e68-04e5-4530-8d94-859789faeb94-kube-api-access-jnrxw\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303277 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-internal-tls-certs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303308 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-scripts\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.303356 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-combined-ca-bundle\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.304785 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68092e68-04e5-4530-8d94-859789faeb94-logs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.314079 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa998ea4-f50d-4441-b6ad-b160a19ea4a9" path="/var/lib/kubelet/pods/aa998ea4-f50d-4441-b6ad-b160a19ea4a9/volumes" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.314902 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-79ccdbbbbd-gl27l"] Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.314927 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r6j6b"] Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.316343 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.328677 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-combined-ca-bundle\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.347889 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-public-tls-certs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.348070 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-internal-tls-certs\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.348139 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-scripts\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.349926 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-config-data\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.356353 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r6j6b"] Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.360769 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnrxw\" (UniqueName: \"kubernetes.io/projected/68092e68-04e5-4530-8d94-859789faeb94-kube-api-access-jnrxw\") pod \"placement-79ccdbbbbd-gl27l\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.430716 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-678c76989b-8x56d"] Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.432163 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.435166 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.454624 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.454815 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.454931 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-d7pv5" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.455034 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515335 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-httpd-config\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515376 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515430 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-combined-ca-bundle\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515452 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgmt8\" (UniqueName: \"kubernetes.io/projected/e34be903-da25-4cdb-9298-2d53fdce0276-kube-api-access-cgmt8\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515472 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515520 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-config\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515541 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515607 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czvqb\" (UniqueName: \"kubernetes.io/projected/f26a6d05-e0ac-4f17-bcd9-fc011996b052-kube-api-access-czvqb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515631 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-ovndb-tls-certs\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515650 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-svc\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.515697 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-config\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.539292 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-678c76989b-8x56d"] Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.616801 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-config\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.616881 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-httpd-config\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.616904 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.616945 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-combined-ca-bundle\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.616965 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgmt8\" (UniqueName: \"kubernetes.io/projected/e34be903-da25-4cdb-9298-2d53fdce0276-kube-api-access-cgmt8\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.616982 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.617160 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-config\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.617220 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.617368 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czvqb\" (UniqueName: \"kubernetes.io/projected/f26a6d05-e0ac-4f17-bcd9-fc011996b052-kube-api-access-czvqb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.617396 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-ovndb-tls-certs\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.617411 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-svc\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.618535 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.618877 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-config\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.618911 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-svc\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.619129 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.619489 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.625482 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-config\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.628142 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-combined-ca-bundle\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.628674 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-ovndb-tls-certs\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.628767 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-httpd-config\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.642274 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgmt8\" (UniqueName: \"kubernetes.io/projected/e34be903-da25-4cdb-9298-2d53fdce0276-kube-api-access-cgmt8\") pod \"neutron-678c76989b-8x56d\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.643141 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czvqb\" (UniqueName: \"kubernetes.io/projected/f26a6d05-e0ac-4f17-bcd9-fc011996b052-kube-api-access-czvqb\") pod \"dnsmasq-dns-55f844cf75-r6j6b\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.828900 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.848622 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.973292 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c2b13038-d271-48f5-bd28-a38e2b9dff02","Type":"ContainerStarted","Data":"cad2d2b9a9ac73ae35a814e1cadf9d57066e520b238036be878f7dfdb34aabb4"} Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.979833 4778 generic.go:334] "Generic (PLEG): container finished" podID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerID="cc6fc61a82e88c3140b3629f45196f98ee08d5f2fdb0df9b40fe66806a0ccbfd" exitCode=0 Mar 12 13:32:52 crc kubenswrapper[4778]: I0312 13:32:52.979876 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" event={"ID":"39bd75fd-958e-4b3b-abd5-860adf376fd7","Type":"ContainerDied","Data":"cc6fc61a82e88c3140b3629f45196f98ee08d5f2fdb0df9b40fe66806a0ccbfd"} Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.558602 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7955c84d65-qfgcn"] Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.560670 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.564998 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.565443 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.599446 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7955c84d65-qfgcn"] Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.662920 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-public-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.662983 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-config\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.663011 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-httpd-config\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.663103 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h89hz\" (UniqueName: \"kubernetes.io/projected/d582b80a-57bd-4cd4-9e72-8a963cae187d-kube-api-access-h89hz\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.663167 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-combined-ca-bundle\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.663223 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-ovndb-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.663256 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-internal-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.764955 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-config\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.765012 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-httpd-config\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.765063 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h89hz\" (UniqueName: \"kubernetes.io/projected/d582b80a-57bd-4cd4-9e72-8a963cae187d-kube-api-access-h89hz\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.765099 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-combined-ca-bundle\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.765133 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-ovndb-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.765149 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-internal-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.765214 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-public-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.771767 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-public-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.772557 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-combined-ca-bundle\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.776990 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-internal-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.777054 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-config\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.778250 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-ovndb-tls-certs\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.782817 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-httpd-config\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.785444 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h89hz\" (UniqueName: \"kubernetes.io/projected/d582b80a-57bd-4cd4-9e72-8a963cae187d-kube-api-access-h89hz\") pod \"neutron-7955c84d65-qfgcn\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:54 crc kubenswrapper[4778]: I0312 13:32:54.902972 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.038544 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-56sfj" event={"ID":"1af573ef-51c3-4bfc-8de6-eb1be8b75c76","Type":"ContainerDied","Data":"ebc59f76f06ba10050abce1212b94660f06fec69cccf436b21f8ae1838a2520b"} Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.038822 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebc59f76f06ba10050abce1212b94660f06fec69cccf436b21f8ae1838a2520b" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.041490 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" event={"ID":"39bd75fd-958e-4b3b-abd5-860adf376fd7","Type":"ContainerDied","Data":"415522e7cc2372bb11dfe09957497d4a3efac5b28086b59aebe2586918e3f99d"} Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.041511 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="415522e7cc2372bb11dfe09957497d4a3efac5b28086b59aebe2586918e3f99d" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.045590 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.052272 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.130256 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-scripts\") pod \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.130312 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-sb\") pod \"39bd75fd-958e-4b3b-abd5-860adf376fd7\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.130347 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-swift-storage-0\") pod \"39bd75fd-958e-4b3b-abd5-860adf376fd7\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132428 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-svc\") pod \"39bd75fd-958e-4b3b-abd5-860adf376fd7\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132510 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-config-data\") pod \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132561 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-credential-keys\") pod \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132591 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7vfh\" (UniqueName: \"kubernetes.io/projected/39bd75fd-958e-4b3b-abd5-860adf376fd7-kube-api-access-p7vfh\") pod \"39bd75fd-958e-4b3b-abd5-860adf376fd7\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132659 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-fernet-keys\") pod \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132711 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-config\") pod \"39bd75fd-958e-4b3b-abd5-860adf376fd7\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132755 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-nb\") pod \"39bd75fd-958e-4b3b-abd5-860adf376fd7\" (UID: \"39bd75fd-958e-4b3b-abd5-860adf376fd7\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132789 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94tqw\" (UniqueName: \"kubernetes.io/projected/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-kube-api-access-94tqw\") pod \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.132815 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-combined-ca-bundle\") pod \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\" (UID: \"1af573ef-51c3-4bfc-8de6-eb1be8b75c76\") " Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.155514 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-kube-api-access-94tqw" (OuterVolumeSpecName: "kube-api-access-94tqw") pod "1af573ef-51c3-4bfc-8de6-eb1be8b75c76" (UID: "1af573ef-51c3-4bfc-8de6-eb1be8b75c76"). InnerVolumeSpecName "kube-api-access-94tqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.156321 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-scripts" (OuterVolumeSpecName: "scripts") pod "1af573ef-51c3-4bfc-8de6-eb1be8b75c76" (UID: "1af573ef-51c3-4bfc-8de6-eb1be8b75c76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.161373 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1af573ef-51c3-4bfc-8de6-eb1be8b75c76" (UID: "1af573ef-51c3-4bfc-8de6-eb1be8b75c76"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.165696 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1af573ef-51c3-4bfc-8de6-eb1be8b75c76" (UID: "1af573ef-51c3-4bfc-8de6-eb1be8b75c76"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.171892 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39bd75fd-958e-4b3b-abd5-860adf376fd7-kube-api-access-p7vfh" (OuterVolumeSpecName: "kube-api-access-p7vfh") pod "39bd75fd-958e-4b3b-abd5-860adf376fd7" (UID: "39bd75fd-958e-4b3b-abd5-860adf376fd7"). InnerVolumeSpecName "kube-api-access-p7vfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.212334 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-config-data" (OuterVolumeSpecName: "config-data") pod "1af573ef-51c3-4bfc-8de6-eb1be8b75c76" (UID: "1af573ef-51c3-4bfc-8de6-eb1be8b75c76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.219397 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1af573ef-51c3-4bfc-8de6-eb1be8b75c76" (UID: "1af573ef-51c3-4bfc-8de6-eb1be8b75c76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.221907 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "39bd75fd-958e-4b3b-abd5-860adf376fd7" (UID: "39bd75fd-958e-4b3b-abd5-860adf376fd7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.222138 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-config" (OuterVolumeSpecName: "config") pod "39bd75fd-958e-4b3b-abd5-860adf376fd7" (UID: "39bd75fd-958e-4b3b-abd5-860adf376fd7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.231576 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "39bd75fd-958e-4b3b-abd5-860adf376fd7" (UID: "39bd75fd-958e-4b3b-abd5-860adf376fd7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.235615 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.235779 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.235863 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.235942 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.236026 4778 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.236108 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7vfh\" (UniqueName: \"kubernetes.io/projected/39bd75fd-958e-4b3b-abd5-860adf376fd7-kube-api-access-p7vfh\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.236199 4778 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.236279 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.236371 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94tqw\" (UniqueName: \"kubernetes.io/projected/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-kube-api-access-94tqw\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.236447 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af573ef-51c3-4bfc-8de6-eb1be8b75c76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.238569 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39bd75fd-958e-4b3b-abd5-860adf376fd7" (UID: "39bd75fd-958e-4b3b-abd5-860adf376fd7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.242494 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "39bd75fd-958e-4b3b-abd5-860adf376fd7" (UID: "39bd75fd-958e-4b3b-abd5-860adf376fd7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.338823 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.338858 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bd75fd-958e-4b3b-abd5-860adf376fd7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.366477 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r6j6b"] Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.439722 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-79ccdbbbbd-gl27l"] Mar 12 13:32:58 crc kubenswrapper[4778]: W0312 13:32:58.446893 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68092e68_04e5_4530_8d94_859789faeb94.slice/crio-6225b0b7ab31929807b7000d1c797565cb38b8453f9487cc91d0a8fcf517ace6 WatchSource:0}: Error finding container 6225b0b7ab31929807b7000d1c797565cb38b8453f9487cc91d0a8fcf517ace6: Status 404 returned error can't find the container with id 6225b0b7ab31929807b7000d1c797565cb38b8453f9487cc91d0a8fcf517ace6 Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.534991 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-678c76989b-8x56d"] Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.557597 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.557658 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:32:58 crc kubenswrapper[4778]: I0312 13:32:58.805536 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7955c84d65-qfgcn"] Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.103171 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac92f5c5-e457-4915-a919-0dbe3df23ce8","Type":"ContainerStarted","Data":"7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.105898 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7955c84d65-qfgcn" event={"ID":"d582b80a-57bd-4cd4-9e72-8a963cae187d","Type":"ContainerStarted","Data":"284482a4b85498fbfd683802fcf5305643f5a4cf33d63effbb2a1f2fd1071a11"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.112348 4778 generic.go:334] "Generic (PLEG): container finished" podID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerID="663c434423b37b8d735c566ad324f30f6c179866c4697ae6a88fd9aeb0c4709a" exitCode=0 Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.112405 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" event={"ID":"f26a6d05-e0ac-4f17-bcd9-fc011996b052","Type":"ContainerDied","Data":"663c434423b37b8d735c566ad324f30f6c179866c4697ae6a88fd9aeb0c4709a"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.112424 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" event={"ID":"f26a6d05-e0ac-4f17-bcd9-fc011996b052","Type":"ContainerStarted","Data":"eedcc18be187ca3b0fbc761493f2664ac917f21e94152db48e6204214d9b050b"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.130640 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.130618976 podStartE2EDuration="10.130618976s" podCreationTimestamp="2026-03-12 13:32:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:59.127205399 +0000 UTC m=+1397.575900795" watchObservedRunningTime="2026-03-12 13:32:59.130618976 +0000 UTC m=+1397.579314372" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.133049 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-678c76989b-8x56d" event={"ID":"e34be903-da25-4cdb-9298-2d53fdce0276","Type":"ContainerStarted","Data":"73ff3b874391ffdc31812d5d85f13741c2920b13dddb21f9bdace835187b0822"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.157494 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79ccdbbbbd-gl27l" event={"ID":"68092e68-04e5-4530-8d94-859789faeb94","Type":"ContainerStarted","Data":"be846a255557e511860dc7bc1b884d65bc6e48bfb1b98ae1316cb74617623c2b"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.157547 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79ccdbbbbd-gl27l" event={"ID":"68092e68-04e5-4530-8d94-859789faeb94","Type":"ContainerStarted","Data":"6225b0b7ab31929807b7000d1c797565cb38b8453f9487cc91d0a8fcf517ace6"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.169150 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-69b6dc4885-6lrlq"] Mar 12 13:32:59 crc kubenswrapper[4778]: E0312 13:32:59.169898 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af573ef-51c3-4bfc-8de6-eb1be8b75c76" containerName="keystone-bootstrap" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.169924 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af573ef-51c3-4bfc-8de6-eb1be8b75c76" containerName="keystone-bootstrap" Mar 12 13:32:59 crc kubenswrapper[4778]: E0312 13:32:59.169960 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="init" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.169969 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="init" Mar 12 13:32:59 crc kubenswrapper[4778]: E0312 13:32:59.169992 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="dnsmasq-dns" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.170000 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="dnsmasq-dns" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.170248 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="dnsmasq-dns" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.170271 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af573ef-51c3-4bfc-8de6-eb1be8b75c76" containerName="keystone-bootstrap" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.171163 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.178922 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.179247 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.184870 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-56sfj" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.185141 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c2b13038-d271-48f5-bd28-a38e2b9dff02","Type":"ContainerStarted","Data":"d321738b43c55df790b0a01418c177d18aaa7772e4cf7fca03bdeedb1c32e127"} Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.185299 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.188472 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-69b6dc4885-6lrlq"] Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.267217 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.26717709 podStartE2EDuration="9.26717709s" podCreationTimestamp="2026-03-12 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:32:59.251951418 +0000 UTC m=+1397.700646824" watchObservedRunningTime="2026-03-12 13:32:59.26717709 +0000 UTC m=+1397.715872486" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285253 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-credential-keys\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285327 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-config-data\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285354 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-public-tls-certs\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285377 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-scripts\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285439 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-internal-tls-certs\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285475 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbnjm\" (UniqueName: \"kubernetes.io/projected/a56bb599-f10d-4564-b6bf-48128dc2c7f1-kube-api-access-bbnjm\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285583 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-fernet-keys\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.285680 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-combined-ca-bundle\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.313747 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-v2vtk"] Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.330317 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-v2vtk"] Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387372 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-combined-ca-bundle\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387447 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-credential-keys\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387478 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-config-data\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387497 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-public-tls-certs\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387513 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-scripts\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387541 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-internal-tls-certs\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387574 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbnjm\" (UniqueName: \"kubernetes.io/projected/a56bb599-f10d-4564-b6bf-48128dc2c7f1-kube-api-access-bbnjm\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.387625 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-fernet-keys\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.392212 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-fernet-keys\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.403632 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-internal-tls-certs\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.404497 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-public-tls-certs\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.404633 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-credential-keys\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.409726 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-scripts\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.411777 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-combined-ca-bundle\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.416594 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a56bb599-f10d-4564-b6bf-48128dc2c7f1-config-data\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.424205 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbnjm\" (UniqueName: \"kubernetes.io/projected/a56bb599-f10d-4564-b6bf-48128dc2c7f1-kube-api-access-bbnjm\") pod \"keystone-69b6dc4885-6lrlq\" (UID: \"a56bb599-f10d-4564-b6bf-48128dc2c7f1\") " pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:32:59 crc kubenswrapper[4778]: I0312 13:32:59.551646 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.027170 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-v2vtk" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.148:5353: i/o timeout" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.061013 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-69b6dc4885-6lrlq"] Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.204208 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79ccdbbbbd-gl27l" event={"ID":"68092e68-04e5-4530-8d94-859789faeb94","Type":"ContainerStarted","Data":"8cdda802eadd8c68b3ba4b5b69b6a0fd021902af043f1083daaae42e4e3ba4bc"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.204595 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.204609 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.216480 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerStarted","Data":"78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.218562 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-69b6dc4885-6lrlq" event={"ID":"a56bb599-f10d-4564-b6bf-48128dc2c7f1","Type":"ContainerStarted","Data":"f2bbb1c1d63926d49e0698557adc54b702c70a83918c69d3788645fc858ad68a"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.220345 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p59s9" event={"ID":"a682334f-73c0-4e38-8f95-e5de661319bb","Type":"ContainerStarted","Data":"5a74043e2f16e3024a4f2ed6f0c9502985ad493a8f1362a42f34265b2e50d313"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.222854 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7955c84d65-qfgcn" event={"ID":"d582b80a-57bd-4cd4-9e72-8a963cae187d","Type":"ContainerStarted","Data":"71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.222897 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7955c84d65-qfgcn" event={"ID":"d582b80a-57bd-4cd4-9e72-8a963cae187d","Type":"ContainerStarted","Data":"938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.223081 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.227281 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-79ccdbbbbd-gl27l" podStartSLOduration=8.227270824 podStartE2EDuration="8.227270824s" podCreationTimestamp="2026-03-12 13:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:00.222667823 +0000 UTC m=+1398.671363249" watchObservedRunningTime="2026-03-12 13:33:00.227270824 +0000 UTC m=+1398.675966220" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.230163 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" event={"ID":"f26a6d05-e0ac-4f17-bcd9-fc011996b052","Type":"ContainerStarted","Data":"601d1f3845ef933f076859b186d3267da0a7df161ebfa472c4f14f7e9cbd4ec0"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.230403 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.235668 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-678c76989b-8x56d" event={"ID":"e34be903-da25-4cdb-9298-2d53fdce0276","Type":"ContainerStarted","Data":"7423051fcfb7c12e56b049e90be94c641f82520ceab5181c7fcca6713588c77f"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.235717 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-678c76989b-8x56d" event={"ID":"e34be903-da25-4cdb-9298-2d53fdce0276","Type":"ContainerStarted","Data":"76d710be6da7b239e82f6228977b9799ccd95f2824b23913a0585897e926dd74"} Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.237574 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.238385 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-p59s9" podStartSLOduration=3.216684908 podStartE2EDuration="40.238373489s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="2026-03-12 13:32:21.924254977 +0000 UTC m=+1360.372950373" lastFinishedPulling="2026-03-12 13:32:58.945943558 +0000 UTC m=+1397.394638954" observedRunningTime="2026-03-12 13:33:00.236793364 +0000 UTC m=+1398.685488760" watchObservedRunningTime="2026-03-12 13:33:00.238373489 +0000 UTC m=+1398.687068885" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.266949 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" podStartSLOduration=8.266930899 podStartE2EDuration="8.266930899s" podCreationTimestamp="2026-03-12 13:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:00.260064924 +0000 UTC m=+1398.708760330" watchObservedRunningTime="2026-03-12 13:33:00.266930899 +0000 UTC m=+1398.715626295" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.275921 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39bd75fd-958e-4b3b-abd5-860adf376fd7" path="/var/lib/kubelet/pods/39bd75fd-958e-4b3b-abd5-860adf376fd7/volumes" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.300014 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-678c76989b-8x56d" podStartSLOduration=8.299998197 podStartE2EDuration="8.299998197s" podCreationTimestamp="2026-03-12 13:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:00.299207855 +0000 UTC m=+1398.747903251" watchObservedRunningTime="2026-03-12 13:33:00.299998197 +0000 UTC m=+1398.748693593" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.322629 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7955c84d65-qfgcn" podStartSLOduration=6.322604088 podStartE2EDuration="6.322604088s" podCreationTimestamp="2026-03-12 13:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:00.318874842 +0000 UTC m=+1398.767570238" watchObservedRunningTime="2026-03-12 13:33:00.322604088 +0000 UTC m=+1398.771299474" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.602338 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.602396 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.640064 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.644315 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.644396 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.661610 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.706509 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:00 crc kubenswrapper[4778]: I0312 13:33:00.710485 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.251448 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-69b6dc4885-6lrlq" event={"ID":"a56bb599-f10d-4564-b6bf-48128dc2c7f1","Type":"ContainerStarted","Data":"c264a6d2fe3e794592b5aa0308d5c7aff717ee4b7a8d492b2d37e4ff699f5b25"} Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.252853 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.252898 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.252918 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.252935 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.252950 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 13:33:01 crc kubenswrapper[4778]: I0312 13:33:01.289352 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-69b6dc4885-6lrlq" podStartSLOduration=2.289329429 podStartE2EDuration="2.289329429s" podCreationTimestamp="2026-03-12 13:32:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:01.280916831 +0000 UTC m=+1399.729612237" watchObservedRunningTime="2026-03-12 13:33:01.289329429 +0000 UTC m=+1399.738024815" Mar 12 13:33:02 crc kubenswrapper[4778]: I0312 13:33:02.264786 4778 generic.go:334] "Generic (PLEG): container finished" podID="a682334f-73c0-4e38-8f95-e5de661319bb" containerID="5a74043e2f16e3024a4f2ed6f0c9502985ad493a8f1362a42f34265b2e50d313" exitCode=0 Mar 12 13:33:02 crc kubenswrapper[4778]: I0312 13:33:02.271333 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p59s9" event={"ID":"a682334f-73c0-4e38-8f95-e5de661319bb","Type":"ContainerDied","Data":"5a74043e2f16e3024a4f2ed6f0c9502985ad493a8f1362a42f34265b2e50d313"} Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.274518 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d5pl9" event={"ID":"bb110a1e-6281-437d-b857-eb79c4953e1a","Type":"ContainerStarted","Data":"4711a6f852c8bf6a8fa62e985008d918b7971ec55784fb38d2f086199f1f3aee"} Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.274556 4778 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.274545 4778 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.298966 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-d5pl9" podStartSLOduration=2.795487831 podStartE2EDuration="43.298948744s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="2026-03-12 13:32:21.516338667 +0000 UTC m=+1359.965034053" lastFinishedPulling="2026-03-12 13:33:02.01979957 +0000 UTC m=+1400.468494966" observedRunningTime="2026-03-12 13:33:03.292950454 +0000 UTC m=+1401.741645870" watchObservedRunningTime="2026-03-12 13:33:03.298948744 +0000 UTC m=+1401.747644160" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.335309 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.459024 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.718227 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p59s9" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.807853 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-db-sync-config-data\") pod \"a682334f-73c0-4e38-8f95-e5de661319bb\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.807988 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-combined-ca-bundle\") pod \"a682334f-73c0-4e38-8f95-e5de661319bb\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.808017 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5hmn\" (UniqueName: \"kubernetes.io/projected/a682334f-73c0-4e38-8f95-e5de661319bb-kube-api-access-r5hmn\") pod \"a682334f-73c0-4e38-8f95-e5de661319bb\" (UID: \"a682334f-73c0-4e38-8f95-e5de661319bb\") " Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.834430 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a682334f-73c0-4e38-8f95-e5de661319bb" (UID: "a682334f-73c0-4e38-8f95-e5de661319bb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.834678 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a682334f-73c0-4e38-8f95-e5de661319bb-kube-api-access-r5hmn" (OuterVolumeSpecName: "kube-api-access-r5hmn") pod "a682334f-73c0-4e38-8f95-e5de661319bb" (UID: "a682334f-73c0-4e38-8f95-e5de661319bb"). InnerVolumeSpecName "kube-api-access-r5hmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.840168 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a682334f-73c0-4e38-8f95-e5de661319bb" (UID: "a682334f-73c0-4e38-8f95-e5de661319bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.918245 4778 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.918284 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a682334f-73c0-4e38-8f95-e5de661319bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:03 crc kubenswrapper[4778]: I0312 13:33:03.918295 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5hmn\" (UniqueName: \"kubernetes.io/projected/a682334f-73c0-4e38-8f95-e5de661319bb-kube-api-access-r5hmn\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.288788 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p59s9" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.288905 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p59s9" event={"ID":"a682334f-73c0-4e38-8f95-e5de661319bb","Type":"ContainerDied","Data":"9752a8239a23597303e4c0af125d25d5be143749ecb830c3912a0cbc8277763f"} Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.288954 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9752a8239a23597303e4c0af125d25d5be143749ecb830c3912a0cbc8277763f" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.522308 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7dcf9787-ngc87"] Mar 12 13:33:04 crc kubenswrapper[4778]: E0312 13:33:04.522665 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a682334f-73c0-4e38-8f95-e5de661319bb" containerName="barbican-db-sync" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.522681 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a682334f-73c0-4e38-8f95-e5de661319bb" containerName="barbican-db-sync" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.522875 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a682334f-73c0-4e38-8f95-e5de661319bb" containerName="barbican-db-sync" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.523728 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.526526 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.526607 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.527279 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-85xbx" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.604248 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7dcf9787-ngc87"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.616957 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-65c9994dfd-xznqh"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.626281 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.628673 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.637241 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-config-data-custom\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.637483 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d505bb59-3c9e-4cfa-891c-c8e0068e2567-logs\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.637666 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-combined-ca-bundle\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.637805 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-config-data\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.637914 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxbbw\" (UniqueName: \"kubernetes.io/projected/d505bb59-3c9e-4cfa-891c-c8e0068e2567-kube-api-access-zxbbw\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.652114 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-65c9994dfd-xznqh"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740330 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-combined-ca-bundle\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740406 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-config-data-custom\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740448 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-combined-ca-bundle\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740484 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-config-data\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740530 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ee1f546-8428-4b23-93e4-b8370fd4224b-logs\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740557 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-config-data\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740581 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxbbw\" (UniqueName: \"kubernetes.io/projected/d505bb59-3c9e-4cfa-891c-c8e0068e2567-kube-api-access-zxbbw\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740632 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-config-data-custom\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740669 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d505bb59-3c9e-4cfa-891c-c8e0068e2567-logs\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.740716 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqrsf\" (UniqueName: \"kubernetes.io/projected/8ee1f546-8428-4b23-93e4-b8370fd4224b-kube-api-access-qqrsf\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.742293 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d505bb59-3c9e-4cfa-891c-c8e0068e2567-logs\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.746207 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-config-data-custom\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.751358 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-config-data\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.756381 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r6j6b"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.756618 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerName="dnsmasq-dns" containerID="cri-o://601d1f3845ef933f076859b186d3267da0a7df161ebfa472c4f14f7e9cbd4ec0" gracePeriod=10 Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.762133 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.780602 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d505bb59-3c9e-4cfa-891c-c8e0068e2567-combined-ca-bundle\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.782383 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pckv7"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.786959 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.805499 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxbbw\" (UniqueName: \"kubernetes.io/projected/d505bb59-3c9e-4cfa-891c-c8e0068e2567-kube-api-access-zxbbw\") pod \"barbican-worker-7dcf9787-ngc87\" (UID: \"d505bb59-3c9e-4cfa-891c-c8e0068e2567\") " pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.832248 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pckv7"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.842434 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ee1f546-8428-4b23-93e4-b8370fd4224b-logs\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.842496 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-config-data\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.842564 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqrsf\" (UniqueName: \"kubernetes.io/projected/8ee1f546-8428-4b23-93e4-b8370fd4224b-kube-api-access-qqrsf\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.842621 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-combined-ca-bundle\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.842645 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-config-data-custom\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.860615 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-config-data-custom\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.861505 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ee1f546-8428-4b23-93e4-b8370fd4224b-logs\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.861532 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7dcf9787-ngc87" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.862343 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-combined-ca-bundle\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.886019 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee1f546-8428-4b23-93e4-b8370fd4224b-config-data\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.915990 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqrsf\" (UniqueName: \"kubernetes.io/projected/8ee1f546-8428-4b23-93e4-b8370fd4224b-kube-api-access-qqrsf\") pod \"barbican-keystone-listener-65c9994dfd-xznqh\" (UID: \"8ee1f546-8428-4b23-93e4-b8370fd4224b\") " pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.937659 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f884f5564-dxzpv"] Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.939743 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.947827 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.947872 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.947905 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8tvn\" (UniqueName: \"kubernetes.io/projected/0c667b0e-f02d-4e71-959f-5d24b702bd73-kube-api-access-j8tvn\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.947965 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-svc\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.948002 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-config\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.948022 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.949428 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.961616 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" Mar 12 13:33:04 crc kubenswrapper[4778]: I0312 13:33:04.971904 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f884f5564-dxzpv"] Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.050104 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-combined-ca-bundle\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.050893 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-svc\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.050929 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czsp7\" (UniqueName: \"kubernetes.io/projected/ef2e3c21-ccc6-4dcc-a476-7393bb481441-kube-api-access-czsp7\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.050968 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.050994 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-config\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.051021 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.051076 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data-custom\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.051122 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef2e3c21-ccc6-4dcc-a476-7393bb481441-logs\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.051158 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.051177 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.051231 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8tvn\" (UniqueName: \"kubernetes.io/projected/0c667b0e-f02d-4e71-959f-5d24b702bd73-kube-api-access-j8tvn\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.052424 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-config\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.052841 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.052963 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.053144 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-svc\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.053425 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.086069 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8tvn\" (UniqueName: \"kubernetes.io/projected/0c667b0e-f02d-4e71-959f-5d24b702bd73-kube-api-access-j8tvn\") pod \"dnsmasq-dns-85ff748b95-pckv7\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.107638 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.153089 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data-custom\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.153150 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef2e3c21-ccc6-4dcc-a476-7393bb481441-logs\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.153246 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-combined-ca-bundle\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.153277 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czsp7\" (UniqueName: \"kubernetes.io/projected/ef2e3c21-ccc6-4dcc-a476-7393bb481441-kube-api-access-czsp7\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.153305 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.154972 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef2e3c21-ccc6-4dcc-a476-7393bb481441-logs\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.162321 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data-custom\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.163244 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.167001 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-combined-ca-bundle\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.176233 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czsp7\" (UniqueName: \"kubernetes.io/projected/ef2e3c21-ccc6-4dcc-a476-7393bb481441-kube-api-access-czsp7\") pod \"barbican-api-5f884f5564-dxzpv\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.421014 4778 generic.go:334] "Generic (PLEG): container finished" podID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerID="601d1f3845ef933f076859b186d3267da0a7df161ebfa472c4f14f7e9cbd4ec0" exitCode=0 Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.421065 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" event={"ID":"f26a6d05-e0ac-4f17-bcd9-fc011996b052","Type":"ContainerDied","Data":"601d1f3845ef933f076859b186d3267da0a7df161ebfa472c4f14f7e9cbd4ec0"} Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.427228 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.715875 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7dcf9787-ngc87"] Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.944343 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:33:05 crc kubenswrapper[4778]: I0312 13:33:05.950121 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-65c9994dfd-xznqh"] Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.047665 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-nb\") pod \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.048418 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czvqb\" (UniqueName: \"kubernetes.io/projected/f26a6d05-e0ac-4f17-bcd9-fc011996b052-kube-api-access-czvqb\") pod \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.048530 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-swift-storage-0\") pod \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.048566 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-config\") pod \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.048626 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-sb\") pod \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.048665 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-svc\") pod \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\" (UID: \"f26a6d05-e0ac-4f17-bcd9-fc011996b052\") " Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.075450 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f26a6d05-e0ac-4f17-bcd9-fc011996b052-kube-api-access-czvqb" (OuterVolumeSpecName: "kube-api-access-czvqb") pod "f26a6d05-e0ac-4f17-bcd9-fc011996b052" (UID: "f26a6d05-e0ac-4f17-bcd9-fc011996b052"). InnerVolumeSpecName "kube-api-access-czvqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.113167 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-config" (OuterVolumeSpecName: "config") pod "f26a6d05-e0ac-4f17-bcd9-fc011996b052" (UID: "f26a6d05-e0ac-4f17-bcd9-fc011996b052"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.120680 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f884f5564-dxzpv"] Mar 12 13:33:06 crc kubenswrapper[4778]: W0312 13:33:06.124597 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef2e3c21_ccc6_4dcc_a476_7393bb481441.slice/crio-91360286d4706715ddbf7b7dd1e71ab18f2b12552f2316ff72136087f9c79c95 WatchSource:0}: Error finding container 91360286d4706715ddbf7b7dd1e71ab18f2b12552f2316ff72136087f9c79c95: Status 404 returned error can't find the container with id 91360286d4706715ddbf7b7dd1e71ab18f2b12552f2316ff72136087f9c79c95 Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.129652 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f26a6d05-e0ac-4f17-bcd9-fc011996b052" (UID: "f26a6d05-e0ac-4f17-bcd9-fc011996b052"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.138923 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f26a6d05-e0ac-4f17-bcd9-fc011996b052" (UID: "f26a6d05-e0ac-4f17-bcd9-fc011996b052"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.140478 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f26a6d05-e0ac-4f17-bcd9-fc011996b052" (UID: "f26a6d05-e0ac-4f17-bcd9-fc011996b052"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.151208 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czvqb\" (UniqueName: \"kubernetes.io/projected/f26a6d05-e0ac-4f17-bcd9-fc011996b052-kube-api-access-czvqb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.151242 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.151255 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.151264 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.151272 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.156365 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f26a6d05-e0ac-4f17-bcd9-fc011996b052" (UID: "f26a6d05-e0ac-4f17-bcd9-fc011996b052"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.169289 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pckv7"] Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.253291 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f26a6d05-e0ac-4f17-bcd9-fc011996b052-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.382286 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.454144 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.454548 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dcf9787-ngc87" event={"ID":"d505bb59-3c9e-4cfa-891c-c8e0068e2567","Type":"ContainerStarted","Data":"50ae9b33b5ff909778b51cf71217aac63884ec464e0459f118e6c5dcbc107d88"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.488671 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f884f5564-dxzpv" event={"ID":"ef2e3c21-ccc6-4dcc-a476-7393bb481441","Type":"ContainerStarted","Data":"555085059a0c8494fcbd31c46657e06bdebc21317a675fa20661619d5dc02586"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.488717 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f884f5564-dxzpv" event={"ID":"ef2e3c21-ccc6-4dcc-a476-7393bb481441","Type":"ContainerStarted","Data":"91360286d4706715ddbf7b7dd1e71ab18f2b12552f2316ff72136087f9c79c95"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.522100 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" event={"ID":"f26a6d05-e0ac-4f17-bcd9-fc011996b052","Type":"ContainerDied","Data":"eedcc18be187ca3b0fbc761493f2664ac917f21e94152db48e6204214d9b050b"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.522151 4778 scope.go:117] "RemoveContainer" containerID="601d1f3845ef933f076859b186d3267da0a7df161ebfa472c4f14f7e9cbd4ec0" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.522350 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r6j6b" Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.525686 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" event={"ID":"0c667b0e-f02d-4e71-959f-5d24b702bd73","Type":"ContainerStarted","Data":"3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.525719 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" event={"ID":"0c667b0e-f02d-4e71-959f-5d24b702bd73","Type":"ContainerStarted","Data":"155ae76e9da40b17013784a015c926481ecd3a26d03501642e4e09f600be7598"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.532066 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" event={"ID":"8ee1f546-8428-4b23-93e4-b8370fd4224b","Type":"ContainerStarted","Data":"2255414430f846d703036cabdb509256e41f80ca7e9c35ed9c2c678cad9afbb7"} Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.606907 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r6j6b"] Mar 12 13:33:06 crc kubenswrapper[4778]: I0312 13:33:06.625520 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r6j6b"] Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.549288 4778 generic.go:334] "Generic (PLEG): container finished" podID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerID="3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484" exitCode=0 Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.549391 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" event={"ID":"0c667b0e-f02d-4e71-959f-5d24b702bd73","Type":"ContainerDied","Data":"3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484"} Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.826226 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-86cb765474-5pq5z"] Mar 12 13:33:07 crc kubenswrapper[4778]: E0312 13:33:07.826597 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerName="dnsmasq-dns" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.826611 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerName="dnsmasq-dns" Mar 12 13:33:07 crc kubenswrapper[4778]: E0312 13:33:07.826621 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerName="init" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.826627 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerName="init" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.826837 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" containerName="dnsmasq-dns" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.828295 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.830791 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.832628 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.852312 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-86cb765474-5pq5z"] Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.899919 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-internal-tls-certs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.899966 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-config-data-custom\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.899988 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-public-tls-certs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.900010 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-config-data\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.900076 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75gk9\" (UniqueName: \"kubernetes.io/projected/6bd172c5-383f-4273-98a5-2c92223dc765-kube-api-access-75gk9\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.900141 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-combined-ca-bundle\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:07 crc kubenswrapper[4778]: I0312 13:33:07.900171 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd172c5-383f-4273-98a5-2c92223dc765-logs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002088 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75gk9\" (UniqueName: \"kubernetes.io/projected/6bd172c5-383f-4273-98a5-2c92223dc765-kube-api-access-75gk9\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002194 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-combined-ca-bundle\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002230 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd172c5-383f-4273-98a5-2c92223dc765-logs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002269 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-internal-tls-certs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002289 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-config-data-custom\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002305 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-public-tls-certs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002326 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-config-data\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.002793 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd172c5-383f-4273-98a5-2c92223dc765-logs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.010872 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-config-data\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.011597 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-combined-ca-bundle\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.012347 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-config-data-custom\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.015711 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-internal-tls-certs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.028657 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75gk9\" (UniqueName: \"kubernetes.io/projected/6bd172c5-383f-4273-98a5-2c92223dc765-kube-api-access-75gk9\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.033823 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd172c5-383f-4273-98a5-2c92223dc765-public-tls-certs\") pod \"barbican-api-86cb765474-5pq5z\" (UID: \"6bd172c5-383f-4273-98a5-2c92223dc765\") " pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.146891 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:08 crc kubenswrapper[4778]: I0312 13:33:08.286326 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f26a6d05-e0ac-4f17-bcd9-fc011996b052" path="/var/lib/kubelet/pods/f26a6d05-e0ac-4f17-bcd9-fc011996b052/volumes" Mar 12 13:33:11 crc kubenswrapper[4778]: I0312 13:33:11.588959 4778 generic.go:334] "Generic (PLEG): container finished" podID="bb110a1e-6281-437d-b857-eb79c4953e1a" containerID="4711a6f852c8bf6a8fa62e985008d918b7971ec55784fb38d2f086199f1f3aee" exitCode=0 Mar 12 13:33:11 crc kubenswrapper[4778]: I0312 13:33:11.589059 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d5pl9" event={"ID":"bb110a1e-6281-437d-b857-eb79c4953e1a","Type":"ContainerDied","Data":"4711a6f852c8bf6a8fa62e985008d918b7971ec55784fb38d2f086199f1f3aee"} Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.144549 4778 scope.go:117] "RemoveContainer" containerID="663c434423b37b8d735c566ad324f30f6c179866c4697ae6a88fd9aeb0c4709a" Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.600557 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f884f5564-dxzpv" event={"ID":"ef2e3c21-ccc6-4dcc-a476-7393bb481441","Type":"ContainerStarted","Data":"c0edf91d21f7ba54f7ae8ead172101f785145fe82241acf1f7236f38396130a9"} Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.600921 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.602251 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": dial tcp 10.217.0.162:9311: connect: connection refused" Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.606545 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" event={"ID":"0c667b0e-f02d-4e71-959f-5d24b702bd73","Type":"ContainerStarted","Data":"12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e"} Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.606672 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.609803 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dcf9787-ngc87" event={"ID":"d505bb59-3c9e-4cfa-891c-c8e0068e2567","Type":"ContainerStarted","Data":"c64ec1f96cc1fcd84def68a1a08ab9a56b79de4d573fa1686270714d0b8a5ddc"} Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.652041 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f884f5564-dxzpv" podStartSLOduration=8.652019412 podStartE2EDuration="8.652019412s" podCreationTimestamp="2026-03-12 13:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:12.623300927 +0000 UTC m=+1411.071996343" watchObservedRunningTime="2026-03-12 13:33:12.652019412 +0000 UTC m=+1411.100714808" Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.653846 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-86cb765474-5pq5z"] Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.657333 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" podStartSLOduration=8.657323222 podStartE2EDuration="8.657323222s" podCreationTimestamp="2026-03-12 13:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:12.642881492 +0000 UTC m=+1411.091576898" watchObservedRunningTime="2026-03-12 13:33:12.657323222 +0000 UTC m=+1411.106018618" Mar 12 13:33:12 crc kubenswrapper[4778]: W0312 13:33:12.739927 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bd172c5_383f_4273_98a5_2c92223dc765.slice/crio-c60e87485b8c433cd0ae87ab204a606d44fb181b69fc47c512c8260657aee02f WatchSource:0}: Error finding container c60e87485b8c433cd0ae87ab204a606d44fb181b69fc47c512c8260657aee02f: Status 404 returned error can't find the container with id c60e87485b8c433cd0ae87ab204a606d44fb181b69fc47c512c8260657aee02f Mar 12 13:33:12 crc kubenswrapper[4778]: I0312 13:33:12.917101 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.047263 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-combined-ca-bundle\") pod \"bb110a1e-6281-437d-b857-eb79c4953e1a\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.047726 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-config-data\") pod \"bb110a1e-6281-437d-b857-eb79c4953e1a\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.047795 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-db-sync-config-data\") pod \"bb110a1e-6281-437d-b857-eb79c4953e1a\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.047849 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-scripts\") pod \"bb110a1e-6281-437d-b857-eb79c4953e1a\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.047884 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb110a1e-6281-437d-b857-eb79c4953e1a-etc-machine-id\") pod \"bb110a1e-6281-437d-b857-eb79c4953e1a\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.047909 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpfhh\" (UniqueName: \"kubernetes.io/projected/bb110a1e-6281-437d-b857-eb79c4953e1a-kube-api-access-jpfhh\") pod \"bb110a1e-6281-437d-b857-eb79c4953e1a\" (UID: \"bb110a1e-6281-437d-b857-eb79c4953e1a\") " Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.048459 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb110a1e-6281-437d-b857-eb79c4953e1a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bb110a1e-6281-437d-b857-eb79c4953e1a" (UID: "bb110a1e-6281-437d-b857-eb79c4953e1a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.052565 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-scripts" (OuterVolumeSpecName: "scripts") pod "bb110a1e-6281-437d-b857-eb79c4953e1a" (UID: "bb110a1e-6281-437d-b857-eb79c4953e1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.052604 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb110a1e-6281-437d-b857-eb79c4953e1a-kube-api-access-jpfhh" (OuterVolumeSpecName: "kube-api-access-jpfhh") pod "bb110a1e-6281-437d-b857-eb79c4953e1a" (UID: "bb110a1e-6281-437d-b857-eb79c4953e1a"). InnerVolumeSpecName "kube-api-access-jpfhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.053739 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bb110a1e-6281-437d-b857-eb79c4953e1a" (UID: "bb110a1e-6281-437d-b857-eb79c4953e1a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.078585 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb110a1e-6281-437d-b857-eb79c4953e1a" (UID: "bb110a1e-6281-437d-b857-eb79c4953e1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.108474 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-config-data" (OuterVolumeSpecName: "config-data") pod "bb110a1e-6281-437d-b857-eb79c4953e1a" (UID: "bb110a1e-6281-437d-b857-eb79c4953e1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.150506 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.150547 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.150556 4778 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.150567 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb110a1e-6281-437d-b857-eb79c4953e1a-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.150577 4778 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb110a1e-6281-437d-b857-eb79c4953e1a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.150585 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpfhh\" (UniqueName: \"kubernetes.io/projected/bb110a1e-6281-437d-b857-eb79c4953e1a-kube-api-access-jpfhh\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.623094 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" event={"ID":"8ee1f546-8428-4b23-93e4-b8370fd4224b","Type":"ContainerStarted","Data":"614728f1eac6b8691120431d6140445a3ccbe27708df146ad3e25503f01c0604"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.623140 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" event={"ID":"8ee1f546-8428-4b23-93e4-b8370fd4224b","Type":"ContainerStarted","Data":"014d4595f50991485d408457e25c7f3fa72ebe29191214a4e0fed5284e25223e"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.625776 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dcf9787-ngc87" event={"ID":"d505bb59-3c9e-4cfa-891c-c8e0068e2567","Type":"ContainerStarted","Data":"27318f7e7450cb0dc3ae3f744d5bacf9c41e8a67f279e8cd3f34ad4fb6cb4753"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.628695 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerStarted","Data":"53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.628880 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-central-agent" containerID="cri-o://a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8" gracePeriod=30 Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.628896 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.628966 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="proxy-httpd" containerID="cri-o://53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce" gracePeriod=30 Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.628994 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-notification-agent" containerID="cri-o://f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0" gracePeriod=30 Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.629092 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="sg-core" containerID="cri-o://78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70" gracePeriod=30 Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.637969 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d5pl9" event={"ID":"bb110a1e-6281-437d-b857-eb79c4953e1a","Type":"ContainerDied","Data":"8d37cd44357eb35c5c4917c8593f7e9902991ee071e5d92e025804bd35c2f76e"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.638011 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d37cd44357eb35c5c4917c8593f7e9902991ee071e5d92e025804bd35c2f76e" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.638064 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d5pl9" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.656376 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-65c9994dfd-xznqh" podStartSLOduration=2.801382524 podStartE2EDuration="9.656361911s" podCreationTimestamp="2026-03-12 13:33:04 +0000 UTC" firstStartedPulling="2026-03-12 13:33:05.97173962 +0000 UTC m=+1404.420435016" lastFinishedPulling="2026-03-12 13:33:12.826719007 +0000 UTC m=+1411.275414403" observedRunningTime="2026-03-12 13:33:13.652749218 +0000 UTC m=+1412.101444614" watchObservedRunningTime="2026-03-12 13:33:13.656361911 +0000 UTC m=+1412.105057307" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.656757 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86cb765474-5pq5z" event={"ID":"6bd172c5-383f-4273-98a5-2c92223dc765","Type":"ContainerStarted","Data":"1dd1a3a407b095e68292977490d4973425a0b88c012b1d56585413eb14cc160f"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.656798 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86cb765474-5pq5z" event={"ID":"6bd172c5-383f-4273-98a5-2c92223dc765","Type":"ContainerStarted","Data":"da3c6322756657010fcf36cc524d53cd418cee529c0e474ed330babf44343e04"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.656808 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86cb765474-5pq5z" event={"ID":"6bd172c5-383f-4273-98a5-2c92223dc765","Type":"ContainerStarted","Data":"c60e87485b8c433cd0ae87ab204a606d44fb181b69fc47c512c8260657aee02f"} Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.656934 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.657633 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.688277 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.121743837 podStartE2EDuration="53.688259966s" podCreationTimestamp="2026-03-12 13:32:20 +0000 UTC" firstStartedPulling="2026-03-12 13:32:21.663396629 +0000 UTC m=+1360.112092025" lastFinishedPulling="2026-03-12 13:33:12.229912758 +0000 UTC m=+1410.678608154" observedRunningTime="2026-03-12 13:33:13.677217832 +0000 UTC m=+1412.125913228" watchObservedRunningTime="2026-03-12 13:33:13.688259966 +0000 UTC m=+1412.136955362" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.697871 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7dcf9787-ngc87" podStartSLOduration=3.265820768 podStartE2EDuration="9.697852718s" podCreationTimestamp="2026-03-12 13:33:04 +0000 UTC" firstStartedPulling="2026-03-12 13:33:05.74083537 +0000 UTC m=+1404.189530766" lastFinishedPulling="2026-03-12 13:33:12.17286732 +0000 UTC m=+1410.621562716" observedRunningTime="2026-03-12 13:33:13.69581368 +0000 UTC m=+1412.144509086" watchObservedRunningTime="2026-03-12 13:33:13.697852718 +0000 UTC m=+1412.146548124" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.734291 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-86cb765474-5pq5z" podStartSLOduration=6.734272561 podStartE2EDuration="6.734272561s" podCreationTimestamp="2026-03-12 13:33:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:13.728093246 +0000 UTC m=+1412.176788652" watchObservedRunningTime="2026-03-12 13:33:13.734272561 +0000 UTC m=+1412.182967957" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.940385 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:13 crc kubenswrapper[4778]: E0312 13:33:13.941096 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb110a1e-6281-437d-b857-eb79c4953e1a" containerName="cinder-db-sync" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.941109 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb110a1e-6281-437d-b857-eb79c4953e1a" containerName="cinder-db-sync" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.941283 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb110a1e-6281-437d-b857-eb79c4953e1a" containerName="cinder-db-sync" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.942159 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.945109 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5pxn8" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.945423 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.945455 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.946003 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 12 13:33:13 crc kubenswrapper[4778]: I0312 13:33:13.959530 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.071304 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.071444 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.071480 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.071524 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.071650 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.073673 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r9bc\" (UniqueName: \"kubernetes.io/projected/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-kube-api-access-6r9bc\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.124667 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pckv7"] Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.124740 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-246x7"] Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.126407 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.177104 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.183262 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.183394 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.183452 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.183618 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r9bc\" (UniqueName: \"kubernetes.io/projected/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-kube-api-access-6r9bc\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.183748 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.183884 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.188261 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.188496 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.194589 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.194904 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.197558 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-246x7"] Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.209785 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r9bc\" (UniqueName: \"kubernetes.io/projected/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-kube-api-access-6r9bc\") pod \"cinder-scheduler-0\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.212648 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.214593 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.216715 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.247859 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293166 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293296 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62md\" (UniqueName: \"kubernetes.io/projected/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-kube-api-access-m62md\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293338 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293374 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-scripts\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293408 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2a8e63-4f63-475d-a03a-3f094c697595-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293535 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293608 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dzr\" (UniqueName: \"kubernetes.io/projected/3a2a8e63-4f63-475d-a03a-3f094c697595-kube-api-access-g8dzr\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293674 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293710 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293800 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-config\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293910 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data-custom\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.293940 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.294020 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2a8e63-4f63-475d-a03a-3f094c697595-logs\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.299697 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395228 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2a8e63-4f63-475d-a03a-3f094c697595-logs\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395287 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395327 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62md\" (UniqueName: \"kubernetes.io/projected/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-kube-api-access-m62md\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395349 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395364 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-scripts\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395394 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2a8e63-4f63-475d-a03a-3f094c697595-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395413 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395433 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dzr\" (UniqueName: \"kubernetes.io/projected/3a2a8e63-4f63-475d-a03a-3f094c697595-kube-api-access-g8dzr\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395455 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395473 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395504 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-config\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395537 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data-custom\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395556 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.395638 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2a8e63-4f63-475d-a03a-3f094c697595-logs\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.397318 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2a8e63-4f63-475d-a03a-3f094c697595-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.399883 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.400861 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.406666 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.410223 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-config\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.411107 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.412364 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.413838 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.420967 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data-custom\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.421451 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-scripts\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.423833 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62md\" (UniqueName: \"kubernetes.io/projected/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-kube-api-access-m62md\") pod \"dnsmasq-dns-5c9776ccc5-246x7\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.428765 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dzr\" (UniqueName: \"kubernetes.io/projected/3a2a8e63-4f63-475d-a03a-3f094c697595-kube-api-access-g8dzr\") pod \"cinder-api-0\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.480339 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.585996 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.686832 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerID="53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce" exitCode=0 Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.686856 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerID="78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70" exitCode=2 Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.686863 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerID="a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8" exitCode=0 Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.686987 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerDied","Data":"53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce"} Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.687022 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerDied","Data":"78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70"} Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.687034 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerDied","Data":"a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8"} Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.687826 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerName="dnsmasq-dns" containerID="cri-o://12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e" gracePeriod=10 Mar 12 13:33:14 crc kubenswrapper[4778]: I0312 13:33:14.688125 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.050773 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-246x7"] Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.071586 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.230272 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:15 crc kubenswrapper[4778]: W0312 13:33:15.268553 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a2a8e63_4f63_475d_a03a_3f094c697595.slice/crio-70300bcacb58d72f86e4ec213d3c35fe83a0cd2426b113dee9b8a89577ab08fd WatchSource:0}: Error finding container 70300bcacb58d72f86e4ec213d3c35fe83a0cd2426b113dee9b8a89577ab08fd: Status 404 returned error can't find the container with id 70300bcacb58d72f86e4ec213d3c35fe83a0cd2426b113dee9b8a89577ab08fd Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.452988 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.514635 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-nb\") pod \"0c667b0e-f02d-4e71-959f-5d24b702bd73\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.514760 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-svc\") pod \"0c667b0e-f02d-4e71-959f-5d24b702bd73\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.514800 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-sb\") pod \"0c667b0e-f02d-4e71-959f-5d24b702bd73\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.514844 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8tvn\" (UniqueName: \"kubernetes.io/projected/0c667b0e-f02d-4e71-959f-5d24b702bd73-kube-api-access-j8tvn\") pod \"0c667b0e-f02d-4e71-959f-5d24b702bd73\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.514902 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-config\") pod \"0c667b0e-f02d-4e71-959f-5d24b702bd73\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.514970 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-swift-storage-0\") pod \"0c667b0e-f02d-4e71-959f-5d24b702bd73\" (UID: \"0c667b0e-f02d-4e71-959f-5d24b702bd73\") " Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.532784 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c667b0e-f02d-4e71-959f-5d24b702bd73-kube-api-access-j8tvn" (OuterVolumeSpecName: "kube-api-access-j8tvn") pod "0c667b0e-f02d-4e71-959f-5d24b702bd73" (UID: "0c667b0e-f02d-4e71-959f-5d24b702bd73"). InnerVolumeSpecName "kube-api-access-j8tvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.600400 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c667b0e-f02d-4e71-959f-5d24b702bd73" (UID: "0c667b0e-f02d-4e71-959f-5d24b702bd73"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.607767 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c667b0e-f02d-4e71-959f-5d24b702bd73" (UID: "0c667b0e-f02d-4e71-959f-5d24b702bd73"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.607855 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c667b0e-f02d-4e71-959f-5d24b702bd73" (UID: "0c667b0e-f02d-4e71-959f-5d24b702bd73"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.611882 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c667b0e-f02d-4e71-959f-5d24b702bd73" (UID: "0c667b0e-f02d-4e71-959f-5d24b702bd73"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.616694 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.616733 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.616745 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.616757 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8tvn\" (UniqueName: \"kubernetes.io/projected/0c667b0e-f02d-4e71-959f-5d24b702bd73-kube-api-access-j8tvn\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.616771 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.620004 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-config" (OuterVolumeSpecName: "config") pod "0c667b0e-f02d-4e71-959f-5d24b702bd73" (UID: "0c667b0e-f02d-4e71-959f-5d24b702bd73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.704590 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1","Type":"ContainerStarted","Data":"64a915f04d0f3e7d3a300bf442920e4fae54d16b184cc83aeb6a0de63549b7fc"} Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.706138 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2a8e63-4f63-475d-a03a-3f094c697595","Type":"ContainerStarted","Data":"70300bcacb58d72f86e4ec213d3c35fe83a0cd2426b113dee9b8a89577ab08fd"} Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.713606 4778 generic.go:334] "Generic (PLEG): container finished" podID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerID="3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31" exitCode=0 Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.713713 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" event={"ID":"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3","Type":"ContainerDied","Data":"3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31"} Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.713768 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" event={"ID":"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3","Type":"ContainerStarted","Data":"60f6f77084cfe6904eb9dc78f60c8b66e7fa89e1a236dd4007f1375a76319d3b"} Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.717789 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c667b0e-f02d-4e71-959f-5d24b702bd73-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.721716 4778 generic.go:334] "Generic (PLEG): container finished" podID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerID="12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e" exitCode=0 Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.721823 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.721823 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" event={"ID":"0c667b0e-f02d-4e71-959f-5d24b702bd73","Type":"ContainerDied","Data":"12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e"} Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.721878 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pckv7" event={"ID":"0c667b0e-f02d-4e71-959f-5d24b702bd73","Type":"ContainerDied","Data":"155ae76e9da40b17013784a015c926481ecd3a26d03501642e4e09f600be7598"} Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.721905 4778 scope.go:117] "RemoveContainer" containerID="12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.776702 4778 scope.go:117] "RemoveContainer" containerID="3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.781360 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pckv7"] Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.790667 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pckv7"] Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.882927 4778 scope.go:117] "RemoveContainer" containerID="12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e" Mar 12 13:33:15 crc kubenswrapper[4778]: E0312 13:33:15.883557 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e\": container with ID starting with 12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e not found: ID does not exist" containerID="12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.883610 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e"} err="failed to get container status \"12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e\": rpc error: code = NotFound desc = could not find container \"12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e\": container with ID starting with 12b6f20e5515eed5ea7b7c17096230648e78a0b77fb023b0f4f649bf2c61cb1e not found: ID does not exist" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.883643 4778 scope.go:117] "RemoveContainer" containerID="3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484" Mar 12 13:33:15 crc kubenswrapper[4778]: E0312 13:33:15.884258 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484\": container with ID starting with 3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484 not found: ID does not exist" containerID="3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484" Mar 12 13:33:15 crc kubenswrapper[4778]: I0312 13:33:15.884514 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484"} err="failed to get container status \"3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484\": rpc error: code = NotFound desc = could not find container \"3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484\": container with ID starting with 3d48ca4ea9f31c5d066a8775b412a93231788bd25815af466613f335e2f60484 not found: ID does not exist" Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.073922 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.282318 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" path="/var/lib/kubelet/pods/0c667b0e-f02d-4e71-959f-5d24b702bd73/volumes" Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.540558 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.737729 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2a8e63-4f63-475d-a03a-3f094c697595","Type":"ContainerStarted","Data":"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970"} Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.740078 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" event={"ID":"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3","Type":"ContainerStarted","Data":"3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549"} Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.740203 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:16 crc kubenswrapper[4778]: I0312 13:33:16.760913 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" podStartSLOduration=2.760894863 podStartE2EDuration="2.760894863s" podCreationTimestamp="2026-03-12 13:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:16.756014515 +0000 UTC m=+1415.204709921" watchObservedRunningTime="2026-03-12 13:33:16.760894863 +0000 UTC m=+1415.209590259" Mar 12 13:33:17 crc kubenswrapper[4778]: I0312 13:33:17.765390 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1","Type":"ContainerStarted","Data":"579bc12eaa8aab0c50eb9ede8c49b9d7ccb94f4d26f7a5f51955978076c57a52"} Mar 12 13:33:17 crc kubenswrapper[4778]: I0312 13:33:17.771760 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2a8e63-4f63-475d-a03a-3f094c697595","Type":"ContainerStarted","Data":"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b"} Mar 12 13:33:17 crc kubenswrapper[4778]: I0312 13:33:17.772527 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api-log" containerID="cri-o://9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970" gracePeriod=30 Mar 12 13:33:17 crc kubenswrapper[4778]: I0312 13:33:17.772916 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 12 13:33:17 crc kubenswrapper[4778]: I0312 13:33:17.773103 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api" containerID="cri-o://b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b" gracePeriod=30 Mar 12 13:33:17 crc kubenswrapper[4778]: I0312 13:33:17.797558 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.797542599 podStartE2EDuration="3.797542599s" podCreationTimestamp="2026-03-12 13:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:17.791365414 +0000 UTC m=+1416.240060810" watchObservedRunningTime="2026-03-12 13:33:17.797542599 +0000 UTC m=+1416.246237995" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.445559 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577439 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2a8e63-4f63-475d-a03a-3f094c697595-etc-machine-id\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577561 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a2a8e63-4f63-475d-a03a-3f094c697595-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577574 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2a8e63-4f63-475d-a03a-3f094c697595-logs\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577645 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-combined-ca-bundle\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577737 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577770 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data-custom\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577837 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-scripts\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.577871 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8dzr\" (UniqueName: \"kubernetes.io/projected/3a2a8e63-4f63-475d-a03a-3f094c697595-kube-api-access-g8dzr\") pod \"3a2a8e63-4f63-475d-a03a-3f094c697595\" (UID: \"3a2a8e63-4f63-475d-a03a-3f094c697595\") " Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.578440 4778 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2a8e63-4f63-475d-a03a-3f094c697595-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.578492 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2a8e63-4f63-475d-a03a-3f094c697595-logs" (OuterVolumeSpecName: "logs") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.595460 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2a8e63-4f63-475d-a03a-3f094c697595-kube-api-access-g8dzr" (OuterVolumeSpecName: "kube-api-access-g8dzr") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "kube-api-access-g8dzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.601433 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.602340 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-scripts" (OuterVolumeSpecName: "scripts") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.626365 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.668335 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data" (OuterVolumeSpecName: "config-data") pod "3a2a8e63-4f63-475d-a03a-3f094c697595" (UID: "3a2a8e63-4f63-475d-a03a-3f094c697595"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.680434 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2a8e63-4f63-475d-a03a-3f094c697595-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.680576 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.680655 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.680711 4778 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.680765 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2a8e63-4f63-475d-a03a-3f094c697595-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.680842 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8dzr\" (UniqueName: \"kubernetes.io/projected/3a2a8e63-4f63-475d-a03a-3f094c697595-kube-api-access-g8dzr\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.780833 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1","Type":"ContainerStarted","Data":"6d003c8be41ca71c54434a7c7a1f3fbe12f00352aa1f46649d39fc04831f2c1f"} Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785745 4778 generic.go:334] "Generic (PLEG): container finished" podID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerID="b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b" exitCode=0 Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785772 4778 generic.go:334] "Generic (PLEG): container finished" podID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerID="9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970" exitCode=143 Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785791 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2a8e63-4f63-475d-a03a-3f094c697595","Type":"ContainerDied","Data":"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b"} Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785816 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2a8e63-4f63-475d-a03a-3f094c697595","Type":"ContainerDied","Data":"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970"} Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785827 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2a8e63-4f63-475d-a03a-3f094c697595","Type":"ContainerDied","Data":"70300bcacb58d72f86e4ec213d3c35fe83a0cd2426b113dee9b8a89577ab08fd"} Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785840 4778 scope.go:117] "RemoveContainer" containerID="b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.785962 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.812035 4778 scope.go:117] "RemoveContainer" containerID="9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.819197 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.927133035 podStartE2EDuration="5.819158858s" podCreationTimestamp="2026-03-12 13:33:13 +0000 UTC" firstStartedPulling="2026-03-12 13:33:15.05554944 +0000 UTC m=+1413.504244836" lastFinishedPulling="2026-03-12 13:33:15.947575263 +0000 UTC m=+1414.396270659" observedRunningTime="2026-03-12 13:33:18.818605323 +0000 UTC m=+1417.267300739" watchObservedRunningTime="2026-03-12 13:33:18.819158858 +0000 UTC m=+1417.267854254" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.842348 4778 scope.go:117] "RemoveContainer" containerID="b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b" Mar 12 13:33:18 crc kubenswrapper[4778]: E0312 13:33:18.845256 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b\": container with ID starting with b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b not found: ID does not exist" containerID="b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.845322 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b"} err="failed to get container status \"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b\": rpc error: code = NotFound desc = could not find container \"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b\": container with ID starting with b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b not found: ID does not exist" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.845354 4778 scope.go:117] "RemoveContainer" containerID="9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970" Mar 12 13:33:18 crc kubenswrapper[4778]: E0312 13:33:18.849645 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970\": container with ID starting with 9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970 not found: ID does not exist" containerID="9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.849694 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970"} err="failed to get container status \"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970\": rpc error: code = NotFound desc = could not find container \"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970\": container with ID starting with 9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970 not found: ID does not exist" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.849725 4778 scope.go:117] "RemoveContainer" containerID="b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.850456 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b"} err="failed to get container status \"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b\": rpc error: code = NotFound desc = could not find container \"b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b\": container with ID starting with b7f952975468ddb5173e7d7ed680242b58ea26be1e25587b9f3f0e647fb5dc2b not found: ID does not exist" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.850482 4778 scope.go:117] "RemoveContainer" containerID="9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.854304 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970"} err="failed to get container status \"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970\": rpc error: code = NotFound desc = could not find container \"9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970\": container with ID starting with 9bd2b44aa17d88e68154cd64ae134f6ebbf2fd2a7b21bc7b54275f7cfd102970 not found: ID does not exist" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.854348 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.864459 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873250 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:18 crc kubenswrapper[4778]: E0312 13:33:18.873657 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerName="dnsmasq-dns" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873673 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerName="dnsmasq-dns" Mar 12 13:33:18 crc kubenswrapper[4778]: E0312 13:33:18.873688 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api-log" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873694 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api-log" Mar 12 13:33:18 crc kubenswrapper[4778]: E0312 13:33:18.873704 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerName="init" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873709 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerName="init" Mar 12 13:33:18 crc kubenswrapper[4778]: E0312 13:33:18.873734 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873739 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873953 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c667b0e-f02d-4e71-959f-5d24b702bd73" containerName="dnsmasq-dns" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873981 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.873996 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" containerName="cinder-api-log" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.874956 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.878651 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.878749 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.878808 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.878941 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986245 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986327 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99f72014-50e8-4dd4-9764-1b2c7d546b30-logs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986351 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f72014-50e8-4dd4-9764-1b2c7d546b30-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986402 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-public-tls-certs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986468 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-config-data\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986503 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986537 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-scripts\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986576 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-config-data-custom\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:18 crc kubenswrapper[4778]: I0312 13:33:18.986594 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdkjf\" (UniqueName: \"kubernetes.io/projected/99f72014-50e8-4dd4-9764-1b2c7d546b30-kube-api-access-xdkjf\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090234 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-scripts\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090524 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-config-data-custom\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090547 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdkjf\" (UniqueName: \"kubernetes.io/projected/99f72014-50e8-4dd4-9764-1b2c7d546b30-kube-api-access-xdkjf\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090579 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090609 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f72014-50e8-4dd4-9764-1b2c7d546b30-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090622 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99f72014-50e8-4dd4-9764-1b2c7d546b30-logs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090665 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-public-tls-certs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090703 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-config-data\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.090734 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.091039 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f72014-50e8-4dd4-9764-1b2c7d546b30-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.093890 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99f72014-50e8-4dd4-9764-1b2c7d546b30-logs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.094018 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-config-data-custom\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.095081 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.095502 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-scripts\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.097341 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.100232 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-config-data\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.102685 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99f72014-50e8-4dd4-9764-1b2c7d546b30-public-tls-certs\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.111710 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdkjf\" (UniqueName: \"kubernetes.io/projected/99f72014-50e8-4dd4-9764-1b2c7d546b30-kube-api-access-xdkjf\") pod \"cinder-api-0\" (UID: \"99f72014-50e8-4dd4-9764-1b2c7d546b30\") " pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.197811 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.300164 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.333107 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395604 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-log-httpd\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395717 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-run-httpd\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395778 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-combined-ca-bundle\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395831 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-scripts\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395848 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-config-data\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395884 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-sg-core-conf-yaml\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.395903 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62wfc\" (UniqueName: \"kubernetes.io/projected/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-kube-api-access-62wfc\") pod \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\" (UID: \"b4cb6d6d-bc05-4809-83a7-5aacda62cc10\") " Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.399934 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.400490 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.409245 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-kube-api-access-62wfc" (OuterVolumeSpecName: "kube-api-access-62wfc") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "kube-api-access-62wfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.409335 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-scripts" (OuterVolumeSpecName: "scripts") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.502797 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.509890 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.509933 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.509943 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.509952 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.509963 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62wfc\" (UniqueName: \"kubernetes.io/projected/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-kube-api-access-62wfc\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.548529 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.564748 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-config-data" (OuterVolumeSpecName: "config-data") pod "b4cb6d6d-bc05-4809-83a7-5aacda62cc10" (UID: "b4cb6d6d-bc05-4809-83a7-5aacda62cc10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.611979 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.612043 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cb6d6d-bc05-4809-83a7-5aacda62cc10-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.741304 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.800898 4778 generic.go:334] "Generic (PLEG): container finished" podID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerID="f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0" exitCode=0 Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.801117 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerDied","Data":"f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0"} Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.801145 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4cb6d6d-bc05-4809-83a7-5aacda62cc10","Type":"ContainerDied","Data":"a7e5a5f0fc47985a7306f104e3261cd746e20017382e7ac550b97742b3f6f6e4"} Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.801163 4778 scope.go:117] "RemoveContainer" containerID="53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.801349 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.816495 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99f72014-50e8-4dd4-9764-1b2c7d546b30","Type":"ContainerStarted","Data":"ffc5edd8376b259d2e817c5933836529d349e1e5587b0e6cb096aa98c0eb7270"} Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.849608 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.857822 4778 scope.go:117] "RemoveContainer" containerID="78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.891554 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917263 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.917620 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-central-agent" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917636 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-central-agent" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.917653 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="proxy-httpd" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917659 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="proxy-httpd" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.917678 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="sg-core" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917684 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="sg-core" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.917705 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-notification-agent" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917711 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-notification-agent" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917857 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-notification-agent" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917874 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="sg-core" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917888 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="proxy-httpd" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.917895 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" containerName="ceilometer-central-agent" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.919729 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.924474 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.924680 4778 scope.go:117] "RemoveContainer" containerID="f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.931124 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.945225 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.974991 4778 scope.go:117] "RemoveContainer" containerID="a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.993282 4778 scope.go:117] "RemoveContainer" containerID="53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.993752 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce\": container with ID starting with 53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce not found: ID does not exist" containerID="53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.993783 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce"} err="failed to get container status \"53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce\": rpc error: code = NotFound desc = could not find container \"53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce\": container with ID starting with 53f309e530cf4b086d2af500ee0b6839a2f6cb85c731eee64b12d5c2ca9527ce not found: ID does not exist" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.993841 4778 scope.go:117] "RemoveContainer" containerID="78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.994326 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70\": container with ID starting with 78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70 not found: ID does not exist" containerID="78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.994347 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70"} err="failed to get container status \"78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70\": rpc error: code = NotFound desc = could not find container \"78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70\": container with ID starting with 78dbc9cf48d678718d746451597636002d39908c130023e24550012d03edab70 not found: ID does not exist" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.994360 4778 scope.go:117] "RemoveContainer" containerID="f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.994572 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0\": container with ID starting with f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0 not found: ID does not exist" containerID="f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.994590 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0"} err="failed to get container status \"f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0\": rpc error: code = NotFound desc = could not find container \"f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0\": container with ID starting with f73331cf93a94b368140f81472e855149bad846c050d72495e3f1fdfaa6cf4d0 not found: ID does not exist" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.994604 4778 scope.go:117] "RemoveContainer" containerID="a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8" Mar 12 13:33:19 crc kubenswrapper[4778]: E0312 13:33:19.994791 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8\": container with ID starting with a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8 not found: ID does not exist" containerID="a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8" Mar 12 13:33:19 crc kubenswrapper[4778]: I0312 13:33:19.994810 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8"} err="failed to get container status \"a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8\": rpc error: code = NotFound desc = could not find container \"a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8\": container with ID starting with a2afa1efaa5e813d9e93bd765e7abf6c5129c2365e3e4d71622e5bbd682b89f8 not found: ID does not exist" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.017058 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031124 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfq6z\" (UniqueName: \"kubernetes.io/projected/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-kube-api-access-rfq6z\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031174 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031221 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031258 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-scripts\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031277 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-log-httpd\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031301 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-run-httpd\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.031332 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-config-data\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132582 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfq6z\" (UniqueName: \"kubernetes.io/projected/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-kube-api-access-rfq6z\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132638 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132672 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132712 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-scripts\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132739 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-log-httpd\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132775 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-run-httpd\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.132810 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-config-data\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.136679 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-run-httpd\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.137780 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-log-httpd\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.139913 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.139941 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.140798 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-config-data\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.141543 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-scripts\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.153400 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfq6z\" (UniqueName: \"kubernetes.io/projected/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-kube-api-access-rfq6z\") pod \"ceilometer-0\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.243552 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.267506 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2a8e63-4f63-475d-a03a-3f094c697595" path="/var/lib/kubelet/pods/3a2a8e63-4f63-475d-a03a-3f094c697595/volumes" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.268332 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4cb6d6d-bc05-4809-83a7-5aacda62cc10" path="/var/lib/kubelet/pods/b4cb6d6d-bc05-4809-83a7-5aacda62cc10/volumes" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.334171 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-86cb765474-5pq5z" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.414442 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f884f5564-dxzpv"] Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.414979 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" containerID="cri-o://555085059a0c8494fcbd31c46657e06bdebc21317a675fa20661619d5dc02586" gracePeriod=30 Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.417087 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" containerID="cri-o://c0edf91d21f7ba54f7ae8ead172101f785145fe82241acf1f7236f38396130a9" gracePeriod=30 Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.428964 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": EOF" Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.849604 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99f72014-50e8-4dd4-9764-1b2c7d546b30","Type":"ContainerStarted","Data":"62fa57c95e565e21b55673b674363abc6eb8ad44ca2bbf998f2181e0eacf6026"} Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.853472 4778 generic.go:334] "Generic (PLEG): container finished" podID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerID="555085059a0c8494fcbd31c46657e06bdebc21317a675fa20661619d5dc02586" exitCode=143 Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.854307 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f884f5564-dxzpv" event={"ID":"ef2e3c21-ccc6-4dcc-a476-7393bb481441","Type":"ContainerDied","Data":"555085059a0c8494fcbd31c46657e06bdebc21317a675fa20661619d5dc02586"} Mar 12 13:33:20 crc kubenswrapper[4778]: I0312 13:33:20.902747 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:20 crc kubenswrapper[4778]: W0312 13:33:20.928877 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b2fa220_02b1_4940_9ae0_3d9e5b4bcd9c.slice/crio-824d71ac269215e859c8ef1b41498f4804c8adec49c2375a8307421f28798e4b WatchSource:0}: Error finding container 824d71ac269215e859c8ef1b41498f4804c8adec49c2375a8307421f28798e4b: Status 404 returned error can't find the container with id 824d71ac269215e859c8ef1b41498f4804c8adec49c2375a8307421f28798e4b Mar 12 13:33:21 crc kubenswrapper[4778]: I0312 13:33:21.867875 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99f72014-50e8-4dd4-9764-1b2c7d546b30","Type":"ContainerStarted","Data":"384c98ecb5fea55607fd01b63ae71a434b6a533af87e4618a66f51cb22530b62"} Mar 12 13:33:21 crc kubenswrapper[4778]: I0312 13:33:21.868181 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 12 13:33:21 crc kubenswrapper[4778]: I0312 13:33:21.872739 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerStarted","Data":"824d71ac269215e859c8ef1b41498f4804c8adec49c2375a8307421f28798e4b"} Mar 12 13:33:22 crc kubenswrapper[4778]: I0312 13:33:22.299880 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.2998594709999995 podStartE2EDuration="4.299859471s" podCreationTimestamp="2026-03-12 13:33:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:21.896224591 +0000 UTC m=+1420.344919997" watchObservedRunningTime="2026-03-12 13:33:22.299859471 +0000 UTC m=+1420.748554867" Mar 12 13:33:22 crc kubenswrapper[4778]: I0312 13:33:22.857981 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:33:22 crc kubenswrapper[4778]: I0312 13:33:22.886634 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerStarted","Data":"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c"} Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.087661 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7955c84d65-qfgcn"] Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.088161 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7955c84d65-qfgcn" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-api" containerID="cri-o://938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd" gracePeriod=30 Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.088941 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7955c84d65-qfgcn" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-httpd" containerID="cri-o://71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09" gracePeriod=30 Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.114785 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7955c84d65-qfgcn" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": EOF" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.117411 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-769c65dfd5-t7d9g"] Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.118849 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.145076 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-769c65dfd5-t7d9g"] Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207205 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-combined-ca-bundle\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207285 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f76t\" (UniqueName: \"kubernetes.io/projected/e3118f8b-6bd2-4fba-8300-114513770916-kube-api-access-4f76t\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207364 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-public-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207415 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-httpd-config\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207449 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-config\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207684 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-internal-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.207848 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313284 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313350 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-combined-ca-bundle\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313383 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f76t\" (UniqueName: \"kubernetes.io/projected/e3118f8b-6bd2-4fba-8300-114513770916-kube-api-access-4f76t\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313420 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-public-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313458 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-httpd-config\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313484 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-config\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.313540 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-internal-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.321167 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-internal-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.325342 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-combined-ca-bundle\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.325985 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-httpd-config\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.326876 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-public-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.329386 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-config\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.339007 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.349074 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f76t\" (UniqueName: \"kubernetes.io/projected/e3118f8b-6bd2-4fba-8300-114513770916-kube-api-access-4f76t\") pod \"neutron-769c65dfd5-t7d9g\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.461333 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.800639 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.805122 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.899225 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerStarted","Data":"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7"} Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.899265 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerStarted","Data":"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e"} Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.902006 4778 generic.go:334] "Generic (PLEG): container finished" podID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerID="71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09" exitCode=0 Mar 12 13:33:23 crc kubenswrapper[4778]: I0312 13:33:23.902097 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7955c84d65-qfgcn" event={"ID":"d582b80a-57bd-4cd4-9e72-8a963cae187d","Type":"ContainerDied","Data":"71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.055930 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d4d765698-l7bjx"] Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.057449 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.068126 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d4d765698-l7bjx"] Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138381 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-public-tls-certs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138421 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-scripts\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138461 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-config-data\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138526 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxx2q\" (UniqueName: \"kubernetes.io/projected/267e7df2-d35c-45c4-af65-e8af31f8f6cf-kube-api-access-xxx2q\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138568 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-internal-tls-certs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138602 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-combined-ca-bundle\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.138712 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/267e7df2-d35c-45c4-af65-e8af31f8f6cf-logs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.198980 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-769c65dfd5-t7d9g"] Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240270 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-public-tls-certs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240332 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-scripts\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240388 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-config-data\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240442 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxx2q\" (UniqueName: \"kubernetes.io/projected/267e7df2-d35c-45c4-af65-e8af31f8f6cf-kube-api-access-xxx2q\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240477 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-internal-tls-certs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240506 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-combined-ca-bundle\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.240620 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/267e7df2-d35c-45c4-af65-e8af31f8f6cf-logs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.241053 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/267e7df2-d35c-45c4-af65-e8af31f8f6cf-logs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.245812 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-internal-tls-certs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.246176 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-combined-ca-bundle\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.247151 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-config-data\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.247701 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-public-tls-certs\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.251489 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267e7df2-d35c-45c4-af65-e8af31f8f6cf-scripts\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.261554 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxx2q\" (UniqueName: \"kubernetes.io/projected/267e7df2-d35c-45c4-af65-e8af31f8f6cf-kube-api-access-xxx2q\") pod \"placement-d4d765698-l7bjx\" (UID: \"267e7df2-d35c-45c4-af65-e8af31f8f6cf\") " pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.482271 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.498799 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.557805 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-56bl9"] Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.558084 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="dnsmasq-dns" containerID="cri-o://512c2c0cf187f0ee46cccf1da3f29d083846818126627409ab7b1bb5fa1ef052" gracePeriod=10 Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.586170 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.644901 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.742719 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861474 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-ovndb-tls-certs\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861529 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-internal-tls-certs\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861592 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-config\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861650 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h89hz\" (UniqueName: \"kubernetes.io/projected/d582b80a-57bd-4cd4-9e72-8a963cae187d-kube-api-access-h89hz\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861748 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-combined-ca-bundle\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861795 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-public-tls-certs\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.861831 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-httpd-config\") pod \"d582b80a-57bd-4cd4-9e72-8a963cae187d\" (UID: \"d582b80a-57bd-4cd4-9e72-8a963cae187d\") " Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.872513 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d582b80a-57bd-4cd4-9e72-8a963cae187d-kube-api-access-h89hz" (OuterVolumeSpecName: "kube-api-access-h89hz") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "kube-api-access-h89hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.878570 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.918948 4778 generic.go:334] "Generic (PLEG): container finished" podID="811bc15c-050c-4d37-a19f-095086748286" containerID="512c2c0cf187f0ee46cccf1da3f29d083846818126627409ab7b1bb5fa1ef052" exitCode=0 Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.919017 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" event={"ID":"811bc15c-050c-4d37-a19f-095086748286","Type":"ContainerDied","Data":"512c2c0cf187f0ee46cccf1da3f29d083846818126627409ab7b1bb5fa1ef052"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.932843 4778 generic.go:334] "Generic (PLEG): container finished" podID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerID="938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd" exitCode=0 Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.932923 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7955c84d65-qfgcn" event={"ID":"d582b80a-57bd-4cd4-9e72-8a963cae187d","Type":"ContainerDied","Data":"938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.932957 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7955c84d65-qfgcn" event={"ID":"d582b80a-57bd-4cd4-9e72-8a963cae187d","Type":"ContainerDied","Data":"284482a4b85498fbfd683802fcf5305643f5a4cf33d63effbb2a1f2fd1071a11"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.932980 4778 scope.go:117] "RemoveContainer" containerID="71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.933137 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7955c84d65-qfgcn" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.946304 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.947916 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="cinder-scheduler" containerID="cri-o://579bc12eaa8aab0c50eb9ede8c49b9d7ccb94f4d26f7a5f51955978076c57a52" gracePeriod=30 Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.949071 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-t7d9g" event={"ID":"e3118f8b-6bd2-4fba-8300-114513770916","Type":"ContainerStarted","Data":"6738f9dd946d748869f4b26f4030a90ea55b7a4599f29ac178ad859657a706f7"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.949105 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.949118 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-t7d9g" event={"ID":"e3118f8b-6bd2-4fba-8300-114513770916","Type":"ContainerStarted","Data":"7559ac32cffd7eca339ac8d8d2f5491100a0167d9ce788c2eb95e805cc071cda"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.949129 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-t7d9g" event={"ID":"e3118f8b-6bd2-4fba-8300-114513770916","Type":"ContainerStarted","Data":"8300c5c0870d3a0dc15fa6bca84b387efeba0222c0e9b918971777a65c2fcb29"} Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.949470 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="probe" containerID="cri-o://6d003c8be41ca71c54434a7c7a1f3fbe12f00352aa1f46649d39fc04831f2c1f" gracePeriod=30 Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.954871 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.964537 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.964568 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.964576 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.964585 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h89hz\" (UniqueName: \"kubernetes.io/projected/d582b80a-57bd-4cd4-9e72-8a963cae187d-kube-api-access-h89hz\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.976825 4778 scope.go:117] "RemoveContainer" containerID="938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd" Mar 12 13:33:24 crc kubenswrapper[4778]: I0312 13:33:24.986842 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:24.999873 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-config" (OuterVolumeSpecName: "config") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.000066 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-769c65dfd5-t7d9g" podStartSLOduration=2.000040923 podStartE2EDuration="2.000040923s" podCreationTimestamp="2026-03-12 13:33:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:24.983676089 +0000 UTC m=+1423.432371485" watchObservedRunningTime="2026-03-12 13:33:25.000040923 +0000 UTC m=+1423.448736339" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.010903 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d582b80a-57bd-4cd4-9e72-8a963cae187d" (UID: "d582b80a-57bd-4cd4-9e72-8a963cae187d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.017648 4778 scope.go:117] "RemoveContainer" containerID="71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09" Mar 12 13:33:25 crc kubenswrapper[4778]: E0312 13:33:25.027358 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09\": container with ID starting with 71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09 not found: ID does not exist" containerID="71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.027400 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09"} err="failed to get container status \"71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09\": rpc error: code = NotFound desc = could not find container \"71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09\": container with ID starting with 71d475b828218d4b5f04543cac9306418884b36e07b75eda675a3ad92ddced09 not found: ID does not exist" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.027423 4778 scope.go:117] "RemoveContainer" containerID="938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd" Mar 12 13:33:25 crc kubenswrapper[4778]: E0312 13:33:25.039294 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd\": container with ID starting with 938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd not found: ID does not exist" containerID="938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.039333 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd"} err="failed to get container status \"938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd\": rpc error: code = NotFound desc = could not find container \"938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd\": container with ID starting with 938c7e0b5c63a6fd5773476e5ae689de9d1155fb4dbd3f7bca4dc6764bc762cd not found: ID does not exist" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.066677 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.067322 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.067334 4778 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d582b80a-57bd-4cd4-9e72-8a963cae187d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.117304 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d4d765698-l7bjx"] Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.357711 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.376227 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7955c84d65-qfgcn"] Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.383336 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7955c84d65-qfgcn"] Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.478686 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsg69\" (UniqueName: \"kubernetes.io/projected/811bc15c-050c-4d37-a19f-095086748286-kube-api-access-bsg69\") pod \"811bc15c-050c-4d37-a19f-095086748286\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.478788 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-svc\") pod \"811bc15c-050c-4d37-a19f-095086748286\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.478820 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-sb\") pod \"811bc15c-050c-4d37-a19f-095086748286\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.478855 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-nb\") pod \"811bc15c-050c-4d37-a19f-095086748286\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.478875 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-swift-storage-0\") pod \"811bc15c-050c-4d37-a19f-095086748286\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.479008 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-config\") pod \"811bc15c-050c-4d37-a19f-095086748286\" (UID: \"811bc15c-050c-4d37-a19f-095086748286\") " Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.487736 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/811bc15c-050c-4d37-a19f-095086748286-kube-api-access-bsg69" (OuterVolumeSpecName: "kube-api-access-bsg69") pod "811bc15c-050c-4d37-a19f-095086748286" (UID: "811bc15c-050c-4d37-a19f-095086748286"). InnerVolumeSpecName "kube-api-access-bsg69". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.514422 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.514473 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.541831 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "811bc15c-050c-4d37-a19f-095086748286" (UID: "811bc15c-050c-4d37-a19f-095086748286"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.552543 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "811bc15c-050c-4d37-a19f-095086748286" (UID: "811bc15c-050c-4d37-a19f-095086748286"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.554772 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-config" (OuterVolumeSpecName: "config") pod "811bc15c-050c-4d37-a19f-095086748286" (UID: "811bc15c-050c-4d37-a19f-095086748286"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.555300 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "811bc15c-050c-4d37-a19f-095086748286" (UID: "811bc15c-050c-4d37-a19f-095086748286"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.570043 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "811bc15c-050c-4d37-a19f-095086748286" (UID: "811bc15c-050c-4d37-a19f-095086748286"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.582432 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.582462 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.582472 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.582483 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsg69\" (UniqueName: \"kubernetes.io/projected/811bc15c-050c-4d37-a19f-095086748286-kube-api-access-bsg69\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.582498 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.582509 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/811bc15c-050c-4d37-a19f-095086748286-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.846659 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:57122->10.217.0.162:9311: read: connection reset by peer" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.847019 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f884f5564-dxzpv" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:57136->10.217.0.162:9311: read: connection reset by peer" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.959866 4778 generic.go:334] "Generic (PLEG): container finished" podID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerID="c0edf91d21f7ba54f7ae8ead172101f785145fe82241acf1f7236f38396130a9" exitCode=0 Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.959909 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f884f5564-dxzpv" event={"ID":"ef2e3c21-ccc6-4dcc-a476-7393bb481441","Type":"ContainerDied","Data":"c0edf91d21f7ba54f7ae8ead172101f785145fe82241acf1f7236f38396130a9"} Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.962196 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" event={"ID":"811bc15c-050c-4d37-a19f-095086748286","Type":"ContainerDied","Data":"9e0eacf82432587cd58359c3985b8def0ae32125ba66b4e86532ed5c793bbd04"} Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.962314 4778 scope.go:117] "RemoveContainer" containerID="512c2c0cf187f0ee46cccf1da3f29d083846818126627409ab7b1bb5fa1ef052" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.962488 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.978312 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d4d765698-l7bjx" event={"ID":"267e7df2-d35c-45c4-af65-e8af31f8f6cf","Type":"ContainerStarted","Data":"492cef66a73d9ef8d4c1ff75ec30e17e6b5471c575d66ec71eaee6ad116d9dba"} Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.978364 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.978378 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d4d765698-l7bjx" event={"ID":"267e7df2-d35c-45c4-af65-e8af31f8f6cf","Type":"ContainerStarted","Data":"de153839f5b74d673c597a2017225c7a54792e9e45e4e6d0c0727c267094029f"} Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.978389 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d4d765698-l7bjx" event={"ID":"267e7df2-d35c-45c4-af65-e8af31f8f6cf","Type":"ContainerStarted","Data":"e035032f2ad7923ebe4ad0e88eba6de6fc1c7f2ecded1d871ece8d6c5cf60deb"} Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.978413 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.985510 4778 generic.go:334] "Generic (PLEG): container finished" podID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerID="6d003c8be41ca71c54434a7c7a1f3fbe12f00352aa1f46649d39fc04831f2c1f" exitCode=0 Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.985598 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1","Type":"ContainerDied","Data":"6d003c8be41ca71c54434a7c7a1f3fbe12f00352aa1f46649d39fc04831f2c1f"} Mar 12 13:33:25 crc kubenswrapper[4778]: I0312 13:33:25.991285 4778 scope.go:117] "RemoveContainer" containerID="52a29e484c375a20ac3f8fc8c2aa037eb3038bed507119d164be5bd117815abc" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.008404 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d4d765698-l7bjx" podStartSLOduration=2.008383186 podStartE2EDuration="2.008383186s" podCreationTimestamp="2026-03-12 13:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:26.004579488 +0000 UTC m=+1424.453274904" watchObservedRunningTime="2026-03-12 13:33:26.008383186 +0000 UTC m=+1424.457078572" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.058921 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-56bl9"] Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.065822 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-56bl9"] Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.264892 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="811bc15c-050c-4d37-a19f-095086748286" path="/var/lib/kubelet/pods/811bc15c-050c-4d37-a19f-095086748286/volumes" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.265669 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" path="/var/lib/kubelet/pods/d582b80a-57bd-4cd4-9e72-8a963cae187d/volumes" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.376496 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.502969 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data-custom\") pod \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.503398 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czsp7\" (UniqueName: \"kubernetes.io/projected/ef2e3c21-ccc6-4dcc-a476-7393bb481441-kube-api-access-czsp7\") pod \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.503492 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data\") pod \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.503525 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef2e3c21-ccc6-4dcc-a476-7393bb481441-logs\") pod \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.503568 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-combined-ca-bundle\") pod \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\" (UID: \"ef2e3c21-ccc6-4dcc-a476-7393bb481441\") " Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.504449 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2e3c21-ccc6-4dcc-a476-7393bb481441-logs" (OuterVolumeSpecName: "logs") pod "ef2e3c21-ccc6-4dcc-a476-7393bb481441" (UID: "ef2e3c21-ccc6-4dcc-a476-7393bb481441"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.510943 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ef2e3c21-ccc6-4dcc-a476-7393bb481441" (UID: "ef2e3c21-ccc6-4dcc-a476-7393bb481441"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.510971 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2e3c21-ccc6-4dcc-a476-7393bb481441-kube-api-access-czsp7" (OuterVolumeSpecName: "kube-api-access-czsp7") pod "ef2e3c21-ccc6-4dcc-a476-7393bb481441" (UID: "ef2e3c21-ccc6-4dcc-a476-7393bb481441"). InnerVolumeSpecName "kube-api-access-czsp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.528164 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef2e3c21-ccc6-4dcc-a476-7393bb481441" (UID: "ef2e3c21-ccc6-4dcc-a476-7393bb481441"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.552420 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data" (OuterVolumeSpecName: "config-data") pod "ef2e3c21-ccc6-4dcc-a476-7393bb481441" (UID: "ef2e3c21-ccc6-4dcc-a476-7393bb481441"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.605587 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czsp7\" (UniqueName: \"kubernetes.io/projected/ef2e3c21-ccc6-4dcc-a476-7393bb481441-kube-api-access-czsp7\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.605634 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.605648 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef2e3c21-ccc6-4dcc-a476-7393bb481441-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.605661 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.605674 4778 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef2e3c21-ccc6-4dcc-a476-7393bb481441-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.997000 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f884f5564-dxzpv" Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.996991 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f884f5564-dxzpv" event={"ID":"ef2e3c21-ccc6-4dcc-a476-7393bb481441","Type":"ContainerDied","Data":"91360286d4706715ddbf7b7dd1e71ab18f2b12552f2316ff72136087f9c79c95"} Mar 12 13:33:26 crc kubenswrapper[4778]: I0312 13:33:26.997226 4778 scope.go:117] "RemoveContainer" containerID="c0edf91d21f7ba54f7ae8ead172101f785145fe82241acf1f7236f38396130a9" Mar 12 13:33:27 crc kubenswrapper[4778]: I0312 13:33:27.001682 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerStarted","Data":"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84"} Mar 12 13:33:27 crc kubenswrapper[4778]: I0312 13:33:27.002082 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:33:27 crc kubenswrapper[4778]: I0312 13:33:27.036003 4778 scope.go:117] "RemoveContainer" containerID="555085059a0c8494fcbd31c46657e06bdebc21317a675fa20661619d5dc02586" Mar 12 13:33:27 crc kubenswrapper[4778]: I0312 13:33:27.037369 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.00158709 podStartE2EDuration="8.037342883s" podCreationTimestamp="2026-03-12 13:33:19 +0000 UTC" firstStartedPulling="2026-03-12 13:33:20.932628618 +0000 UTC m=+1419.381324014" lastFinishedPulling="2026-03-12 13:33:25.968384411 +0000 UTC m=+1424.417079807" observedRunningTime="2026-03-12 13:33:27.026844805 +0000 UTC m=+1425.475540201" watchObservedRunningTime="2026-03-12 13:33:27.037342883 +0000 UTC m=+1425.486038309" Mar 12 13:33:27 crc kubenswrapper[4778]: I0312 13:33:27.097966 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f884f5564-dxzpv"] Mar 12 13:33:27 crc kubenswrapper[4778]: I0312 13:33:27.106743 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f884f5564-dxzpv"] Mar 12 13:33:28 crc kubenswrapper[4778]: I0312 13:33:28.268988 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" path="/var/lib/kubelet/pods/ef2e3c21-ccc6-4dcc-a476-7393bb481441/volumes" Mar 12 13:33:28 crc kubenswrapper[4778]: I0312 13:33:28.558003 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:33:28 crc kubenswrapper[4778]: I0312 13:33:28.558097 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:33:28 crc kubenswrapper[4778]: I0312 13:33:28.558217 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:33:28 crc kubenswrapper[4778]: I0312 13:33:28.559289 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"572aad6c3b1a3f7c9ef45b8b4feb0d367e7e7916d0ab8dd064e2b8ac87268c51"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:33:28 crc kubenswrapper[4778]: I0312 13:33:28.559410 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://572aad6c3b1a3f7c9ef45b8b4feb0d367e7e7916d0ab8dd064e2b8ac87268c51" gracePeriod=600 Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.024457 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="572aad6c3b1a3f7c9ef45b8b4feb0d367e7e7916d0ab8dd064e2b8ac87268c51" exitCode=0 Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.024549 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"572aad6c3b1a3f7c9ef45b8b4feb0d367e7e7916d0ab8dd064e2b8ac87268c51"} Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.024935 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e"} Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.024958 4778 scope.go:117] "RemoveContainer" containerID="3b4b372cac8f288fc2585670d5ab7c00c41331f173130d39b164aa74e4e3e398" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.026774 4778 generic.go:334] "Generic (PLEG): container finished" podID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerID="579bc12eaa8aab0c50eb9ede8c49b9d7ccb94f4d26f7a5f51955978076c57a52" exitCode=0 Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.026814 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1","Type":"ContainerDied","Data":"579bc12eaa8aab0c50eb9ede8c49b9d7ccb94f4d26f7a5f51955978076c57a52"} Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.263139 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365256 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-etc-machine-id\") pod \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365604 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data\") pod \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365425 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" (UID: "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365640 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data-custom\") pod \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365802 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-combined-ca-bundle\") pod \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365931 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r9bc\" (UniqueName: \"kubernetes.io/projected/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-kube-api-access-6r9bc\") pod \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.365962 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-scripts\") pod \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\" (UID: \"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1\") " Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.366403 4778 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.371591 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" (UID: "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.371606 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-kube-api-access-6r9bc" (OuterVolumeSpecName: "kube-api-access-6r9bc") pod "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" (UID: "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1"). InnerVolumeSpecName "kube-api-access-6r9bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.371605 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-scripts" (OuterVolumeSpecName: "scripts") pod "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" (UID: "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.415916 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" (UID: "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.467834 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.468052 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r9bc\" (UniqueName: \"kubernetes.io/projected/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-kube-api-access-6r9bc\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.468144 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.468218 4778 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.475935 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data" (OuterVolumeSpecName: "config-data") pod "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" (UID: "a3d67881-ce3f-4abe-b07b-a0b22f1f53d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:29 crc kubenswrapper[4778]: I0312 13:33:29.569773 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.038292 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3d67881-ce3f-4abe-b07b-a0b22f1f53d1","Type":"ContainerDied","Data":"64a915f04d0f3e7d3a300bf442920e4fae54d16b184cc83aeb6a0de63549b7fc"} Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.038347 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.038551 4778 scope.go:117] "RemoveContainer" containerID="6d003c8be41ca71c54434a7c7a1f3fbe12f00352aa1f46649d39fc04831f2c1f" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.068779 4778 scope.go:117] "RemoveContainer" containerID="579bc12eaa8aab0c50eb9ede8c49b9d7ccb94f4d26f7a5f51955978076c57a52" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.086400 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.094146 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.119648 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120025 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-httpd" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120037 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-httpd" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120054 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="init" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120062 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="init" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120068 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120076 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120084 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="cinder-scheduler" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120090 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="cinder-scheduler" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120101 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-api" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120107 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-api" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120134 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="dnsmasq-dns" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120139 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="dnsmasq-dns" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120149 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120155 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" Mar 12 13:33:30 crc kubenswrapper[4778]: E0312 13:33:30.120164 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="probe" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120170 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="probe" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120321 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-httpd" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120332 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api-log" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120342 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d582b80a-57bd-4cd4-9e72-8a963cae187d" containerName="neutron-api" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120351 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="dnsmasq-dns" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120368 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2e3c21-ccc6-4dcc-a476-7393bb481441" containerName="barbican-api" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120378 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="cinder-scheduler" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.120385 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" containerName="probe" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.121229 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.125985 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.146835 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.268209 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3d67881-ce3f-4abe-b07b-a0b22f1f53d1" path="/var/lib/kubelet/pods/a3d67881-ce3f-4abe-b07b-a0b22f1f53d1/volumes" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.284673 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/39ee2404-53a8-4598-8c4b-c3a34fbf3480-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.284770 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-scripts\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.284795 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28r84\" (UniqueName: \"kubernetes.io/projected/39ee2404-53a8-4598-8c4b-c3a34fbf3480-kube-api-access-28r84\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.284821 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.285007 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-config-data\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.285054 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.306144 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-56bl9" podUID="811bc15c-050c-4d37-a19f-095086748286" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.386732 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.386886 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/39ee2404-53a8-4598-8c4b-c3a34fbf3480-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.386931 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-scripts\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.386948 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28r84\" (UniqueName: \"kubernetes.io/projected/39ee2404-53a8-4598-8c4b-c3a34fbf3480-kube-api-access-28r84\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.386967 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.387010 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-config-data\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.387549 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/39ee2404-53a8-4598-8c4b-c3a34fbf3480-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.391950 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.392679 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-config-data\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.407139 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-scripts\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.408078 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ee2404-53a8-4598-8c4b-c3a34fbf3480-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.416638 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28r84\" (UniqueName: \"kubernetes.io/projected/39ee2404-53a8-4598-8c4b-c3a34fbf3480-kube-api-access-28r84\") pod \"cinder-scheduler-0\" (UID: \"39ee2404-53a8-4598-8c4b-c3a34fbf3480\") " pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.459943 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 13:33:30 crc kubenswrapper[4778]: I0312 13:33:30.950826 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 13:33:31 crc kubenswrapper[4778]: I0312 13:33:31.056028 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"39ee2404-53a8-4598-8c4b-c3a34fbf3480","Type":"ContainerStarted","Data":"9369f3c2bdb81c52320468f6a0a675db90bbc2790b389d857f47ec6c2986b217"} Mar 12 13:33:31 crc kubenswrapper[4778]: I0312 13:33:31.336583 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 12 13:33:31 crc kubenswrapper[4778]: I0312 13:33:31.463587 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-69b6dc4885-6lrlq" Mar 12 13:33:32 crc kubenswrapper[4778]: I0312 13:33:32.070947 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"39ee2404-53a8-4598-8c4b-c3a34fbf3480","Type":"ContainerStarted","Data":"5caf7b2dd6a6e32cbb2140761105fd053f1b58a4ef8c454f45f75eb7d2e8e8e9"} Mar 12 13:33:33 crc kubenswrapper[4778]: I0312 13:33:33.086011 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"39ee2404-53a8-4598-8c4b-c3a34fbf3480","Type":"ContainerStarted","Data":"bffbd49096f6cfd6157d754703da0c11466db663a3d4974d2922be039c3c1c55"} Mar 12 13:33:33 crc kubenswrapper[4778]: I0312 13:33:33.123573 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.123550524 podStartE2EDuration="3.123550524s" podCreationTimestamp="2026-03-12 13:33:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:33.119154529 +0000 UTC m=+1431.567849925" watchObservedRunningTime="2026-03-12 13:33:33.123550524 +0000 UTC m=+1431.572245930" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.882725 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-77f887c49f-fw2qd"] Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.884368 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.887304 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.891746 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.904556 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.914301 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-77f887c49f-fw2qd"] Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998070 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbd76cb8-462f-4e60-b755-ef3170e70d11-log-httpd\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998323 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-public-tls-certs\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998421 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbd76cb8-462f-4e60-b755-ef3170e70d11-run-httpd\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998494 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-config-data\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998552 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-combined-ca-bundle\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998582 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bbd76cb8-462f-4e60-b755-ef3170e70d11-etc-swift\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998623 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-internal-tls-certs\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:34 crc kubenswrapper[4778]: I0312 13:33:34.998656 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptf49\" (UniqueName: \"kubernetes.io/projected/bbd76cb8-462f-4e60-b755-ef3170e70d11-kube-api-access-ptf49\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.099981 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-combined-ca-bundle\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100039 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bbd76cb8-462f-4e60-b755-ef3170e70d11-etc-swift\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100091 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-internal-tls-certs\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100122 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptf49\" (UniqueName: \"kubernetes.io/projected/bbd76cb8-462f-4e60-b755-ef3170e70d11-kube-api-access-ptf49\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100157 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbd76cb8-462f-4e60-b755-ef3170e70d11-log-httpd\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100225 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-public-tls-certs\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100268 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbd76cb8-462f-4e60-b755-ef3170e70d11-run-httpd\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100328 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-config-data\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.100887 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbd76cb8-462f-4e60-b755-ef3170e70d11-run-httpd\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.101081 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbd76cb8-462f-4e60-b755-ef3170e70d11-log-httpd\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.106127 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-combined-ca-bundle\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.106534 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-config-data\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.107203 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-internal-tls-certs\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.108479 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbd76cb8-462f-4e60-b755-ef3170e70d11-public-tls-certs\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.109805 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bbd76cb8-462f-4e60-b755-ef3170e70d11-etc-swift\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.127041 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptf49\" (UniqueName: \"kubernetes.io/projected/bbd76cb8-462f-4e60-b755-ef3170e70d11-kube-api-access-ptf49\") pod \"swift-proxy-77f887c49f-fw2qd\" (UID: \"bbd76cb8-462f-4e60-b755-ef3170e70d11\") " pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.202048 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.460298 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 12 13:33:35 crc kubenswrapper[4778]: I0312 13:33:35.765745 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-77f887c49f-fw2qd"] Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.112573 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77f887c49f-fw2qd" event={"ID":"bbd76cb8-462f-4e60-b755-ef3170e70d11","Type":"ContainerStarted","Data":"054545cc67ddd38768251482093252364fbb2d00a10b98246b4c1ad92100dd4b"} Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.113967 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77f887c49f-fw2qd" event={"ID":"bbd76cb8-462f-4e60-b755-ef3170e70d11","Type":"ContainerStarted","Data":"d69d23f632a141289fa25126779fdfc5f0cf08505643eae471d81045ee93350c"} Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.184666 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.186038 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.188040 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-lmqm5" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.188276 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.192287 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.205237 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.321410 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/856cd6d1-db21-4503-94d7-cbf27ca96cc2-openstack-config-secret\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.321848 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd5n6\" (UniqueName: \"kubernetes.io/projected/856cd6d1-db21-4503-94d7-cbf27ca96cc2-kube-api-access-jd5n6\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.322007 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856cd6d1-db21-4503-94d7-cbf27ca96cc2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.322315 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/856cd6d1-db21-4503-94d7-cbf27ca96cc2-openstack-config\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.424213 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856cd6d1-db21-4503-94d7-cbf27ca96cc2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.424301 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/856cd6d1-db21-4503-94d7-cbf27ca96cc2-openstack-config\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.425379 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/856cd6d1-db21-4503-94d7-cbf27ca96cc2-openstack-config\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.425461 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/856cd6d1-db21-4503-94d7-cbf27ca96cc2-openstack-config-secret\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.425889 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd5n6\" (UniqueName: \"kubernetes.io/projected/856cd6d1-db21-4503-94d7-cbf27ca96cc2-kube-api-access-jd5n6\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.429634 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/856cd6d1-db21-4503-94d7-cbf27ca96cc2-openstack-config-secret\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.435846 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856cd6d1-db21-4503-94d7-cbf27ca96cc2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.458846 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd5n6\" (UniqueName: \"kubernetes.io/projected/856cd6d1-db21-4503-94d7-cbf27ca96cc2-kube-api-access-jd5n6\") pod \"openstackclient\" (UID: \"856cd6d1-db21-4503-94d7-cbf27ca96cc2\") " pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.507638 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.926591 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.927822 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-central-agent" containerID="cri-o://92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c" gracePeriod=30 Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.927950 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="proxy-httpd" containerID="cri-o://cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84" gracePeriod=30 Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.927994 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="sg-core" containerID="cri-o://e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7" gracePeriod=30 Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.928029 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-notification-agent" containerID="cri-o://a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e" gracePeriod=30 Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.934623 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 12 13:33:36 crc kubenswrapper[4778]: I0312 13:33:36.993210 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.002097 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.122721 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"856cd6d1-db21-4503-94d7-cbf27ca96cc2","Type":"ContainerStarted","Data":"5fee156d13d3b3eafc7b482a7d9e15c3c8309e71dc17da054fde272c1e944ca3"} Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.124849 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77f887c49f-fw2qd" event={"ID":"bbd76cb8-462f-4e60-b755-ef3170e70d11","Type":"ContainerStarted","Data":"af03b29d9375025c0539ee0addb00329a847129df3c98ac51efce5d92af6fdd7"} Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.124915 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.124930 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.127741 4778 generic.go:334] "Generic (PLEG): container finished" podID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerID="cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84" exitCode=0 Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.127770 4778 generic.go:334] "Generic (PLEG): container finished" podID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerID="e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7" exitCode=2 Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.127790 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerDied","Data":"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84"} Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.127813 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerDied","Data":"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7"} Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.147916 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-77f887c49f-fw2qd" podStartSLOduration=3.147895426 podStartE2EDuration="3.147895426s" podCreationTimestamp="2026-03-12 13:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:37.143756359 +0000 UTC m=+1435.592451765" watchObservedRunningTime="2026-03-12 13:33:37.147895426 +0000 UTC m=+1435.596590822" Mar 12 13:33:37 crc kubenswrapper[4778]: I0312 13:33:37.925587 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.057404 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-log-httpd\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.057474 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-sg-core-conf-yaml\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.057652 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-scripts\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.057701 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-run-httpd\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.058153 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.058219 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-config-data\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.058216 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.058314 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfq6z\" (UniqueName: \"kubernetes.io/projected/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-kube-api-access-rfq6z\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.058731 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-combined-ca-bundle\") pod \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\" (UID: \"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c\") " Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.059396 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.059415 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.067309 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-scripts" (OuterVolumeSpecName: "scripts") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.067406 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-kube-api-access-rfq6z" (OuterVolumeSpecName: "kube-api-access-rfq6z") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "kube-api-access-rfq6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.091278 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.159261 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.160968 4778 generic.go:334] "Generic (PLEG): container finished" podID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerID="a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e" exitCode=0 Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.161006 4778 generic.go:334] "Generic (PLEG): container finished" podID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerID="92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c" exitCode=0 Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.161143 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerDied","Data":"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e"} Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.161267 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerDied","Data":"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c"} Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.161353 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c","Type":"ContainerDied","Data":"824d71ac269215e859c8ef1b41498f4804c8adec49c2375a8307421f28798e4b"} Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.161331 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.161317 4778 scope.go:117] "RemoveContainer" containerID="cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.162656 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.162683 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfq6z\" (UniqueName: \"kubernetes.io/projected/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-kube-api-access-rfq6z\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.162698 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.162711 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.236172 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-config-data" (OuterVolumeSpecName: "config-data") pod "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" (UID: "7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.264328 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.323833 4778 scope.go:117] "RemoveContainer" containerID="e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.344576 4778 scope.go:117] "RemoveContainer" containerID="a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.366799 4778 scope.go:117] "RemoveContainer" containerID="92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.385038 4778 scope.go:117] "RemoveContainer" containerID="cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.385697 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84\": container with ID starting with cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84 not found: ID does not exist" containerID="cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.385790 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84"} err="failed to get container status \"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84\": rpc error: code = NotFound desc = could not find container \"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84\": container with ID starting with cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84 not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.385869 4778 scope.go:117] "RemoveContainer" containerID="e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.386224 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7\": container with ID starting with e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7 not found: ID does not exist" containerID="e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.386277 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7"} err="failed to get container status \"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7\": rpc error: code = NotFound desc = could not find container \"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7\": container with ID starting with e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7 not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.386311 4778 scope.go:117] "RemoveContainer" containerID="a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.386645 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e\": container with ID starting with a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e not found: ID does not exist" containerID="a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.386729 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e"} err="failed to get container status \"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e\": rpc error: code = NotFound desc = could not find container \"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e\": container with ID starting with a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.386805 4778 scope.go:117] "RemoveContainer" containerID="92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.387122 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c\": container with ID starting with 92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c not found: ID does not exist" containerID="92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.387165 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c"} err="failed to get container status \"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c\": rpc error: code = NotFound desc = could not find container \"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c\": container with ID starting with 92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.387208 4778 scope.go:117] "RemoveContainer" containerID="cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.387509 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84"} err="failed to get container status \"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84\": rpc error: code = NotFound desc = could not find container \"cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84\": container with ID starting with cda3d15fbae3fde616e8dc2f2ce28f309b9b98d6a94b912a86966cec07509e84 not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.387531 4778 scope.go:117] "RemoveContainer" containerID="e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.387870 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7"} err="failed to get container status \"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7\": rpc error: code = NotFound desc = could not find container \"e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7\": container with ID starting with e22d50b8fe9f90a6aab9adf00774ed799ec453df3b0b299a334bf282330ef1b7 not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.387900 4778 scope.go:117] "RemoveContainer" containerID="a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.388206 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e"} err="failed to get container status \"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e\": rpc error: code = NotFound desc = could not find container \"a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e\": container with ID starting with a1d99c14b9faebc510bf1668c9726ad77d4df3734c20a3bc0e28ff53683f982e not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.388230 4778 scope.go:117] "RemoveContainer" containerID="92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.388540 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c"} err="failed to get container status \"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c\": rpc error: code = NotFound desc = could not find container \"92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c\": container with ID starting with 92af10636577795c46a7d7213efc323d787b43d9aee552320b8e29e2d94b148c not found: ID does not exist" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.491348 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.498923 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.515525 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.515896 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="proxy-httpd" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.515912 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="proxy-httpd" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.515926 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-notification-agent" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.515933 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-notification-agent" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.515947 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-central-agent" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.515954 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-central-agent" Mar 12 13:33:38 crc kubenswrapper[4778]: E0312 13:33:38.515966 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="sg-core" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.515972 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="sg-core" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.516176 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="proxy-httpd" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.516209 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="sg-core" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.516220 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-central-agent" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.516237 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" containerName="ceilometer-notification-agent" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.517734 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.519633 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.519841 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.531048 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.670983 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-run-httpd\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.671130 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.671160 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.671435 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-config-data\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.671495 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-log-httpd\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.671615 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvbsh\" (UniqueName: \"kubernetes.io/projected/76973a17-7173-486f-af83-14c0378fa581-kube-api-access-lvbsh\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.671658 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-scripts\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773118 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvbsh\" (UniqueName: \"kubernetes.io/projected/76973a17-7173-486f-af83-14c0378fa581-kube-api-access-lvbsh\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773169 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-scripts\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773247 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-run-httpd\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773293 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773316 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773509 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-config-data\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.773558 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-log-httpd\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.774196 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-run-httpd\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.774429 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-log-httpd\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.778255 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.778398 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-config-data\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.779387 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.792370 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-scripts\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.800394 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvbsh\" (UniqueName: \"kubernetes.io/projected/76973a17-7173-486f-af83-14c0378fa581-kube-api-access-lvbsh\") pod \"ceilometer-0\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " pod="openstack/ceilometer-0" Mar 12 13:33:38 crc kubenswrapper[4778]: I0312 13:33:38.848149 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:39 crc kubenswrapper[4778]: I0312 13:33:39.308252 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:39 crc kubenswrapper[4778]: W0312 13:33:39.319570 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76973a17_7173_486f_af83_14c0378fa581.slice/crio-0c674ab196e6e3ca8f4c50f5d0e86fbac274fa04d05f2445f7220ac78d3eb924 WatchSource:0}: Error finding container 0c674ab196e6e3ca8f4c50f5d0e86fbac274fa04d05f2445f7220ac78d3eb924: Status 404 returned error can't find the container with id 0c674ab196e6e3ca8f4c50f5d0e86fbac274fa04d05f2445f7220ac78d3eb924 Mar 12 13:33:40 crc kubenswrapper[4778]: I0312 13:33:40.192161 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerStarted","Data":"0c674ab196e6e3ca8f4c50f5d0e86fbac274fa04d05f2445f7220ac78d3eb924"} Mar 12 13:33:40 crc kubenswrapper[4778]: I0312 13:33:40.212448 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:40 crc kubenswrapper[4778]: I0312 13:33:40.277651 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c" path="/var/lib/kubelet/pods/7b2fa220-02b1-4940-9ae0-3d9e5b4bcd9c/volumes" Mar 12 13:33:40 crc kubenswrapper[4778]: I0312 13:33:40.677491 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 12 13:33:41 crc kubenswrapper[4778]: I0312 13:33:41.203073 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerStarted","Data":"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79"} Mar 12 13:33:41 crc kubenswrapper[4778]: I0312 13:33:41.203608 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerStarted","Data":"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a"} Mar 12 13:33:44 crc kubenswrapper[4778]: I0312 13:33:44.692733 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:45 crc kubenswrapper[4778]: I0312 13:33:45.214540 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-77f887c49f-fw2qd" Mar 12 13:33:48 crc kubenswrapper[4778]: I0312 13:33:48.279895 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"856cd6d1-db21-4503-94d7-cbf27ca96cc2","Type":"ContainerStarted","Data":"483cf45dad2053f4851724913051cb9f082ec5ef16f52defcc123d6751b912b0"} Mar 12 13:33:48 crc kubenswrapper[4778]: I0312 13:33:48.293287 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerStarted","Data":"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1"} Mar 12 13:33:48 crc kubenswrapper[4778]: I0312 13:33:48.305099 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.439150789 podStartE2EDuration="12.305082559s" podCreationTimestamp="2026-03-12 13:33:36 +0000 UTC" firstStartedPulling="2026-03-12 13:33:37.001907866 +0000 UTC m=+1435.450603262" lastFinishedPulling="2026-03-12 13:33:47.867839626 +0000 UTC m=+1446.316535032" observedRunningTime="2026-03-12 13:33:48.298121972 +0000 UTC m=+1446.746817378" watchObservedRunningTime="2026-03-12 13:33:48.305082559 +0000 UTC m=+1446.753777955" Mar 12 13:33:49 crc kubenswrapper[4778]: I0312 13:33:49.916466 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:33:49 crc kubenswrapper[4778]: I0312 13:33:49.917418 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-httpd" containerID="cri-o://d321738b43c55df790b0a01418c177d18aaa7772e4cf7fca03bdeedb1c32e127" gracePeriod=30 Mar 12 13:33:49 crc kubenswrapper[4778]: I0312 13:33:49.922196 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-log" containerID="cri-o://cad2d2b9a9ac73ae35a814e1cadf9d57066e520b238036be878f7dfdb34aabb4" gracePeriod=30 Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.313742 4778 generic.go:334] "Generic (PLEG): container finished" podID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerID="cad2d2b9a9ac73ae35a814e1cadf9d57066e520b238036be878f7dfdb34aabb4" exitCode=143 Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.313816 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c2b13038-d271-48f5-bd28-a38e2b9dff02","Type":"ContainerDied","Data":"cad2d2b9a9ac73ae35a814e1cadf9d57066e520b238036be878f7dfdb34aabb4"} Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.316860 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerStarted","Data":"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028"} Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.317018 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-central-agent" containerID="cri-o://a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" gracePeriod=30 Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.317039 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.317062 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="sg-core" containerID="cri-o://9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" gracePeriod=30 Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.317091 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="proxy-httpd" containerID="cri-o://1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" gracePeriod=30 Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.317111 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-notification-agent" containerID="cri-o://7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" gracePeriod=30 Mar 12 13:33:50 crc kubenswrapper[4778]: I0312 13:33:50.338638 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.983690607 podStartE2EDuration="12.338590782s" podCreationTimestamp="2026-03-12 13:33:38 +0000 UTC" firstStartedPulling="2026-03-12 13:33:39.322125761 +0000 UTC m=+1437.770821157" lastFinishedPulling="2026-03-12 13:33:49.677025936 +0000 UTC m=+1448.125721332" observedRunningTime="2026-03-12 13:33:50.335140664 +0000 UTC m=+1448.783836060" watchObservedRunningTime="2026-03-12 13:33:50.338590782 +0000 UTC m=+1448.787286188" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.252466 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318274 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-combined-ca-bundle\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318331 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-sg-core-conf-yaml\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318403 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-scripts\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318434 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-config-data\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318469 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvbsh\" (UniqueName: \"kubernetes.io/projected/76973a17-7173-486f-af83-14c0378fa581-kube-api-access-lvbsh\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318528 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-log-httpd\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.318580 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-run-httpd\") pod \"76973a17-7173-486f-af83-14c0378fa581\" (UID: \"76973a17-7173-486f-af83-14c0378fa581\") " Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.319083 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.319102 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.320482 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.320890 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76973a17-7173-486f-af83-14c0378fa581-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.326395 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76973a17-7173-486f-af83-14c0378fa581-kube-api-access-lvbsh" (OuterVolumeSpecName: "kube-api-access-lvbsh") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "kube-api-access-lvbsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.326531 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-scripts" (OuterVolumeSpecName: "scripts") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.335868 4778 generic.go:334] "Generic (PLEG): container finished" podID="76973a17-7173-486f-af83-14c0378fa581" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" exitCode=0 Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.335901 4778 generic.go:334] "Generic (PLEG): container finished" podID="76973a17-7173-486f-af83-14c0378fa581" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" exitCode=2 Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.335911 4778 generic.go:334] "Generic (PLEG): container finished" podID="76973a17-7173-486f-af83-14c0378fa581" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" exitCode=0 Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.335938 4778 generic.go:334] "Generic (PLEG): container finished" podID="76973a17-7173-486f-af83-14c0378fa581" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" exitCode=0 Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.335960 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerDied","Data":"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028"} Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.335987 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerDied","Data":"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1"} Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.336016 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerDied","Data":"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79"} Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.336028 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerDied","Data":"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a"} Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.336036 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76973a17-7173-486f-af83-14c0378fa581","Type":"ContainerDied","Data":"0c674ab196e6e3ca8f4c50f5d0e86fbac274fa04d05f2445f7220ac78d3eb924"} Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.336053 4778 scope.go:117] "RemoveContainer" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.336233 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.350307 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.372613 4778 scope.go:117] "RemoveContainer" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.392163 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.397427 4778 scope.go:117] "RemoveContainer" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.422331 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.422372 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.422387 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.422399 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvbsh\" (UniqueName: \"kubernetes.io/projected/76973a17-7173-486f-af83-14c0378fa581-kube-api-access-lvbsh\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.435723 4778 scope.go:117] "RemoveContainer" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.440143 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-config-data" (OuterVolumeSpecName: "config-data") pod "76973a17-7173-486f-af83-14c0378fa581" (UID: "76973a17-7173-486f-af83-14c0378fa581"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.460636 4778 scope.go:117] "RemoveContainer" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.461206 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": container with ID starting with 1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028 not found: ID does not exist" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.461273 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028"} err="failed to get container status \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": rpc error: code = NotFound desc = could not find container \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": container with ID starting with 1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.461307 4778 scope.go:117] "RemoveContainer" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.461816 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": container with ID starting with 9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1 not found: ID does not exist" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.461848 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1"} err="failed to get container status \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": rpc error: code = NotFound desc = could not find container \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": container with ID starting with 9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.461865 4778 scope.go:117] "RemoveContainer" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.462254 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": container with ID starting with 7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79 not found: ID does not exist" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.462313 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79"} err="failed to get container status \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": rpc error: code = NotFound desc = could not find container \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": container with ID starting with 7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.462350 4778 scope.go:117] "RemoveContainer" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.462633 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": container with ID starting with a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a not found: ID does not exist" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.462664 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a"} err="failed to get container status \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": rpc error: code = NotFound desc = could not find container \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": container with ID starting with a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.462683 4778 scope.go:117] "RemoveContainer" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.463294 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028"} err="failed to get container status \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": rpc error: code = NotFound desc = could not find container \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": container with ID starting with 1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.463326 4778 scope.go:117] "RemoveContainer" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.463641 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1"} err="failed to get container status \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": rpc error: code = NotFound desc = could not find container \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": container with ID starting with 9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.463667 4778 scope.go:117] "RemoveContainer" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464060 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79"} err="failed to get container status \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": rpc error: code = NotFound desc = could not find container \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": container with ID starting with 7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464082 4778 scope.go:117] "RemoveContainer" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464398 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a"} err="failed to get container status \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": rpc error: code = NotFound desc = could not find container \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": container with ID starting with a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464424 4778 scope.go:117] "RemoveContainer" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464670 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028"} err="failed to get container status \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": rpc error: code = NotFound desc = could not find container \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": container with ID starting with 1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464695 4778 scope.go:117] "RemoveContainer" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464963 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1"} err="failed to get container status \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": rpc error: code = NotFound desc = could not find container \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": container with ID starting with 9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.464988 4778 scope.go:117] "RemoveContainer" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.465294 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79"} err="failed to get container status \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": rpc error: code = NotFound desc = could not find container \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": container with ID starting with 7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.465316 4778 scope.go:117] "RemoveContainer" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.465553 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a"} err="failed to get container status \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": rpc error: code = NotFound desc = could not find container \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": container with ID starting with a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.465574 4778 scope.go:117] "RemoveContainer" containerID="1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.465815 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028"} err="failed to get container status \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": rpc error: code = NotFound desc = could not find container \"1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028\": container with ID starting with 1a8e0d7e1f3aab5c5624ca9847d187e08f63f5617d5b4835b74a5019bd409028 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.465837 4778 scope.go:117] "RemoveContainer" containerID="9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.466090 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1"} err="failed to get container status \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": rpc error: code = NotFound desc = could not find container \"9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1\": container with ID starting with 9c68d01c3251c2dcefdd1f4d92409f7a771f6097f592a70c23676aaa51a086c1 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.466113 4778 scope.go:117] "RemoveContainer" containerID="7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.466383 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79"} err="failed to get container status \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": rpc error: code = NotFound desc = could not find container \"7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79\": container with ID starting with 7f745be7f683445e4da01145bbe31250962c059abf58eb460e2f1333fb155a79 not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.466406 4778 scope.go:117] "RemoveContainer" containerID="a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.466698 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a"} err="failed to get container status \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": rpc error: code = NotFound desc = could not find container \"a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a\": container with ID starting with a5fd58e9033319df013906b8f30fe6475ac6a783b70a23d68bd819715773d59a not found: ID does not exist" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.524625 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76973a17-7173-486f-af83-14c0378fa581-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.676486 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.696423 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.729629 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.730040 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-central-agent" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730062 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-central-agent" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.730089 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-notification-agent" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730099 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-notification-agent" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.730115 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="proxy-httpd" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730124 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="proxy-httpd" Mar 12 13:33:51 crc kubenswrapper[4778]: E0312 13:33:51.730155 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="sg-core" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730163 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="sg-core" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730394 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="sg-core" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730411 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-central-agent" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730431 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="proxy-httpd" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.730445 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76973a17-7173-486f-af83-14c0378fa581" containerName="ceilometer-notification-agent" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.732372 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.746195 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.746643 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.746963 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830248 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830308 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830503 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-run-httpd\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830564 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-scripts\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-config-data\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830878 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-log-httpd\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.830978 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrpfb\" (UniqueName: \"kubernetes.io/projected/0bf8c182-c9d5-4011-b28c-c4f557a8071c-kube-api-access-nrpfb\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.933403 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrpfb\" (UniqueName: \"kubernetes.io/projected/0bf8c182-c9d5-4011-b28c-c4f557a8071c-kube-api-access-nrpfb\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.933841 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.933877 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.933937 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-run-httpd\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.933959 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-scripts\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.934015 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-config-data\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.934055 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-log-httpd\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.934636 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-log-httpd\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.934898 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-run-httpd\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.938042 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-scripts\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.938740 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-config-data\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.939988 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.941876 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:51 crc kubenswrapper[4778]: I0312 13:33:51.951415 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrpfb\" (UniqueName: \"kubernetes.io/projected/0bf8c182-c9d5-4011-b28c-c4f557a8071c-kube-api-access-nrpfb\") pod \"ceilometer-0\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " pod="openstack/ceilometer-0" Mar 12 13:33:52 crc kubenswrapper[4778]: I0312 13:33:52.088099 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:33:52 crc kubenswrapper[4778]: I0312 13:33:52.272297 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76973a17-7173-486f-af83-14c0378fa581" path="/var/lib/kubelet/pods/76973a17-7173-486f-af83-14c0378fa581/volumes" Mar 12 13:33:52 crc kubenswrapper[4778]: I0312 13:33:52.604304 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.067899 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:38626->10.217.0.153:9292: read: connection reset by peer" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.067953 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:38636->10.217.0.153:9292: read: connection reset by peer" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.362170 4778 generic.go:334] "Generic (PLEG): container finished" podID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerID="d321738b43c55df790b0a01418c177d18aaa7772e4cf7fca03bdeedb1c32e127" exitCode=0 Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.362490 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c2b13038-d271-48f5-bd28-a38e2b9dff02","Type":"ContainerDied","Data":"d321738b43c55df790b0a01418c177d18aaa7772e4cf7fca03bdeedb1c32e127"} Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.363567 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerStarted","Data":"0f3aa121caf2c1a6a7f5f32c4c791af4c518cf20357d26f2062f2e017c408468"} Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.477947 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.557679 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-678c76989b-8x56d"] Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.558046 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-678c76989b-8x56d" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-httpd" containerID="cri-o://7423051fcfb7c12e56b049e90be94c641f82520ceab5181c7fcca6713588c77f" gracePeriod=30 Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.558277 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-678c76989b-8x56d" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-api" containerID="cri-o://76d710be6da7b239e82f6228977b9799ccd95f2824b23913a0585897e926dd74" gracePeriod=30 Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.659535 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781678 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-logs\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781726 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-httpd-run\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781779 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-combined-ca-bundle\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781805 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfwzz\" (UniqueName: \"kubernetes.io/projected/c2b13038-d271-48f5-bd28-a38e2b9dff02-kube-api-access-xfwzz\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781906 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-scripts\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781966 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-config-data\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.781991 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-internal-tls-certs\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.782020 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c2b13038-d271-48f5-bd28-a38e2b9dff02\" (UID: \"c2b13038-d271-48f5-bd28-a38e2b9dff02\") " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.783503 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.783930 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-logs" (OuterVolumeSpecName: "logs") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.789872 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.790107 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2b13038-d271-48f5-bd28-a38e2b9dff02-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.795040 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b13038-d271-48f5-bd28-a38e2b9dff02-kube-api-access-xfwzz" (OuterVolumeSpecName: "kube-api-access-xfwzz") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "kube-api-access-xfwzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.795148 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.815347 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-scripts" (OuterVolumeSpecName: "scripts") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.822946 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.864030 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.892804 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.892835 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.892865 4778 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.892875 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.892885 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfwzz\" (UniqueName: \"kubernetes.io/projected/c2b13038-d271-48f5-bd28-a38e2b9dff02-kube-api-access-xfwzz\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.904025 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-config-data" (OuterVolumeSpecName: "config-data") pod "c2b13038-d271-48f5-bd28-a38e2b9dff02" (UID: "c2b13038-d271-48f5-bd28-a38e2b9dff02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.916340 4778 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.994698 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b13038-d271-48f5-bd28-a38e2b9dff02-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:53 crc kubenswrapper[4778]: I0312 13:33:53.994729 4778 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.375917 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c2b13038-d271-48f5-bd28-a38e2b9dff02","Type":"ContainerDied","Data":"15eaae81b5ec94e32bcb75db667617fbe51c32c5f0cac153a8a191ff89576b97"} Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.375962 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.377283 4778 scope.go:117] "RemoveContainer" containerID="d321738b43c55df790b0a01418c177d18aaa7772e4cf7fca03bdeedb1c32e127" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.378525 4778 generic.go:334] "Generic (PLEG): container finished" podID="e34be903-da25-4cdb-9298-2d53fdce0276" containerID="7423051fcfb7c12e56b049e90be94c641f82520ceab5181c7fcca6713588c77f" exitCode=0 Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.378629 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-678c76989b-8x56d" event={"ID":"e34be903-da25-4cdb-9298-2d53fdce0276","Type":"ContainerDied","Data":"7423051fcfb7c12e56b049e90be94c641f82520ceab5181c7fcca6713588c77f"} Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.382080 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerStarted","Data":"8581110fc2e8206867368b8c4ae7af28cb79d5341dbf6b92ea91def7d2e28eb6"} Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.382124 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerStarted","Data":"e925c9c4c7aa08744211c517c124058aade623d45fd2e02df2777b4f2df794b2"} Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.406122 4778 scope.go:117] "RemoveContainer" containerID="cad2d2b9a9ac73ae35a814e1cadf9d57066e520b238036be878f7dfdb34aabb4" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.418964 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.438434 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.452510 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:33:54 crc kubenswrapper[4778]: E0312 13:33:54.453069 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-log" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.453138 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-log" Mar 12 13:33:54 crc kubenswrapper[4778]: E0312 13:33:54.453269 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-httpd" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.453324 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-httpd" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.453543 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-log" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.453618 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" containerName="glance-httpd" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.454574 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.457396 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.457967 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.479556 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.605364 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.605676 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.605811 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.605929 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.606026 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa757af-1c91-4b93-8916-5bbd99b8522e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.606129 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48c68\" (UniqueName: \"kubernetes.io/projected/7fa757af-1c91-4b93-8916-5bbd99b8522e-kube-api-access-48c68\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.606270 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fa757af-1c91-4b93-8916-5bbd99b8522e-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.606430 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709379 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709482 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709539 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709588 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709613 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa757af-1c91-4b93-8916-5bbd99b8522e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709655 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48c68\" (UniqueName: \"kubernetes.io/projected/7fa757af-1c91-4b93-8916-5bbd99b8522e-kube-api-access-48c68\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709719 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fa757af-1c91-4b93-8916-5bbd99b8522e-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.709735 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.710113 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.710438 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fa757af-1c91-4b93-8916-5bbd99b8522e-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.710553 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fa757af-1c91-4b93-8916-5bbd99b8522e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.717751 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.718904 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.720214 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.720432 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fa757af-1c91-4b93-8916-5bbd99b8522e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.752938 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48c68\" (UniqueName: \"kubernetes.io/projected/7fa757af-1c91-4b93-8916-5bbd99b8522e-kube-api-access-48c68\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.757098 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fa757af-1c91-4b93-8916-5bbd99b8522e\") " pod="openstack/glance-default-internal-api-0" Mar 12 13:33:54 crc kubenswrapper[4778]: I0312 13:33:54.772085 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 13:33:55 crc kubenswrapper[4778]: I0312 13:33:55.392996 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerStarted","Data":"86ce4b4705307dac2aa00c6fc4314d927b33960ed30dc0799f79715a9adfcdf9"} Mar 12 13:33:55 crc kubenswrapper[4778]: I0312 13:33:55.471330 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 13:33:55 crc kubenswrapper[4778]: I0312 13:33:55.975982 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.268274 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b13038-d271-48f5-bd28-a38e2b9dff02" path="/var/lib/kubelet/pods/c2b13038-d271-48f5-bd28-a38e2b9dff02/volumes" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.334154 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d4d765698-l7bjx" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.421034 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fa757af-1c91-4b93-8916-5bbd99b8522e","Type":"ContainerStarted","Data":"423ab013fd652a6661ad67d357330604f44bea50f0c07b9d7091f54614bcf3b5"} Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.421100 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fa757af-1c91-4b93-8916-5bbd99b8522e","Type":"ContainerStarted","Data":"3e099ffe3a3d1c8b6878dca7cbc40faffead2664c2494336300caa2723a6629b"} Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.484403 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-sckbb"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.486932 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.549423 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-79ccdbbbbd-gl27l"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.565796 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-79ccdbbbbd-gl27l" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-log" containerID="cri-o://be846a255557e511860dc7bc1b884d65bc6e48bfb1b98ae1316cb74617623c2b" gracePeriod=30 Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.566106 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-79ccdbbbbd-gl27l" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-api" containerID="cri-o://8cdda802eadd8c68b3ba4b5b69b6a0fd021902af043f1083daaae42e4e3ba4bc" gracePeriod=30 Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.609843 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sckbb"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.657904 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln84f\" (UniqueName: \"kubernetes.io/projected/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-kube-api-access-ln84f\") pod \"nova-api-db-create-sckbb\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.657975 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-operator-scripts\") pod \"nova-api-db-create-sckbb\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.704993 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-dcf9-account-create-update-2rmjd"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.706278 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.718751 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.736132 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-dcf9-account-create-update-2rmjd"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.759996 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ab681f-51c2-4723-b5b6-58c841185455-operator-scripts\") pod \"nova-api-dcf9-account-create-update-2rmjd\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.760119 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-operator-scripts\") pod \"nova-api-db-create-sckbb\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.760148 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln84f\" (UniqueName: \"kubernetes.io/projected/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-kube-api-access-ln84f\") pod \"nova-api-db-create-sckbb\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.760177 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvxrm\" (UniqueName: \"kubernetes.io/projected/20ab681f-51c2-4723-b5b6-58c841185455-kube-api-access-hvxrm\") pod \"nova-api-dcf9-account-create-update-2rmjd\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.761157 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-operator-scripts\") pod \"nova-api-db-create-sckbb\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.823796 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln84f\" (UniqueName: \"kubernetes.io/projected/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-kube-api-access-ln84f\") pod \"nova-api-db-create-sckbb\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.855869 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sckbb" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.870624 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-x8nht"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.871878 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.876100 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvxrm\" (UniqueName: \"kubernetes.io/projected/20ab681f-51c2-4723-b5b6-58c841185455-kube-api-access-hvxrm\") pod \"nova-api-dcf9-account-create-update-2rmjd\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.876289 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ab681f-51c2-4723-b5b6-58c841185455-operator-scripts\") pod \"nova-api-dcf9-account-create-update-2rmjd\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.877214 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ab681f-51c2-4723-b5b6-58c841185455-operator-scripts\") pod \"nova-api-dcf9-account-create-update-2rmjd\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.932833 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x8nht"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.947308 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvxrm\" (UniqueName: \"kubernetes.io/projected/20ab681f-51c2-4723-b5b6-58c841185455-kube-api-access-hvxrm\") pod \"nova-api-dcf9-account-create-update-2rmjd\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.961280 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2dh9w"] Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.978901 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwnk2\" (UniqueName: \"kubernetes.io/projected/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-kube-api-access-zwnk2\") pod \"nova-cell0-db-create-x8nht\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.979040 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-operator-scripts\") pod \"nova-cell0-db-create-x8nht\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:56 crc kubenswrapper[4778]: I0312 13:33:56.982804 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.030481 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2dh9w"] Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.074515 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7d0f-account-create-update-t2rrl"] Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.076308 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.082335 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.090302 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klzfm\" (UniqueName: \"kubernetes.io/projected/068c02bc-1daf-4029-84f9-39a395d5de3e-kube-api-access-klzfm\") pod \"nova-cell1-db-create-2dh9w\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.090501 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068c02bc-1daf-4029-84f9-39a395d5de3e-operator-scripts\") pod \"nova-cell1-db-create-2dh9w\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.090771 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwnk2\" (UniqueName: \"kubernetes.io/projected/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-kube-api-access-zwnk2\") pod \"nova-cell0-db-create-x8nht\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.090931 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-operator-scripts\") pod \"nova-cell0-db-create-x8nht\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.092143 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-operator-scripts\") pod \"nova-cell0-db-create-x8nht\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.103952 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7d0f-account-create-update-t2rrl"] Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.133808 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwnk2\" (UniqueName: \"kubernetes.io/projected/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-kube-api-access-zwnk2\") pod \"nova-cell0-db-create-x8nht\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.146174 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.193814 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klzfm\" (UniqueName: \"kubernetes.io/projected/068c02bc-1daf-4029-84f9-39a395d5de3e-kube-api-access-klzfm\") pod \"nova-cell1-db-create-2dh9w\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.193882 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092c3556-0255-4e2f-b2c7-e22b8a3d8418-operator-scripts\") pod \"nova-cell0-7d0f-account-create-update-t2rrl\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.193950 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068c02bc-1daf-4029-84f9-39a395d5de3e-operator-scripts\") pod \"nova-cell1-db-create-2dh9w\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.193986 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfdqt\" (UniqueName: \"kubernetes.io/projected/092c3556-0255-4e2f-b2c7-e22b8a3d8418-kube-api-access-nfdqt\") pod \"nova-cell0-7d0f-account-create-update-t2rrl\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.195754 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068c02bc-1daf-4029-84f9-39a395d5de3e-operator-scripts\") pod \"nova-cell1-db-create-2dh9w\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.220731 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klzfm\" (UniqueName: \"kubernetes.io/projected/068c02bc-1daf-4029-84f9-39a395d5de3e-kube-api-access-klzfm\") pod \"nova-cell1-db-create-2dh9w\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.223290 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.294235 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-94ac-account-create-update-rxvgg"] Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.302501 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.305284 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.312133 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-94ac-account-create-update-rxvgg"] Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.313150 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092c3556-0255-4e2f-b2c7-e22b8a3d8418-operator-scripts\") pod \"nova-cell0-7d0f-account-create-update-t2rrl\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.313868 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdqt\" (UniqueName: \"kubernetes.io/projected/092c3556-0255-4e2f-b2c7-e22b8a3d8418-kube-api-access-nfdqt\") pod \"nova-cell0-7d0f-account-create-update-t2rrl\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.321490 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092c3556-0255-4e2f-b2c7-e22b8a3d8418-operator-scripts\") pod \"nova-cell0-7d0f-account-create-update-t2rrl\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.346573 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.354161 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfdqt\" (UniqueName: \"kubernetes.io/projected/092c3556-0255-4e2f-b2c7-e22b8a3d8418-kube-api-access-nfdqt\") pod \"nova-cell0-7d0f-account-create-update-t2rrl\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.417643 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbbc5fa-b903-4296-a3af-75524920938d-operator-scripts\") pod \"nova-cell1-94ac-account-create-update-rxvgg\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.417686 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn6jh\" (UniqueName: \"kubernetes.io/projected/9dbbc5fa-b903-4296-a3af-75524920938d-kube-api-access-rn6jh\") pod \"nova-cell1-94ac-account-create-update-rxvgg\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.443684 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.466486 4778 generic.go:334] "Generic (PLEG): container finished" podID="68092e68-04e5-4530-8d94-859789faeb94" containerID="be846a255557e511860dc7bc1b884d65bc6e48bfb1b98ae1316cb74617623c2b" exitCode=143 Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.466548 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79ccdbbbbd-gl27l" event={"ID":"68092e68-04e5-4530-8d94-859789faeb94","Type":"ContainerDied","Data":"be846a255557e511860dc7bc1b884d65bc6e48bfb1b98ae1316cb74617623c2b"} Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.496397 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerStarted","Data":"013c8995ca90639ba33078e18954ed308111d321639179db05cd00d19ef56702"} Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.496874 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.521436 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbbc5fa-b903-4296-a3af-75524920938d-operator-scripts\") pod \"nova-cell1-94ac-account-create-update-rxvgg\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.521501 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn6jh\" (UniqueName: \"kubernetes.io/projected/9dbbc5fa-b903-4296-a3af-75524920938d-kube-api-access-rn6jh\") pod \"nova-cell1-94ac-account-create-update-rxvgg\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.522156 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbbc5fa-b903-4296-a3af-75524920938d-operator-scripts\") pod \"nova-cell1-94ac-account-create-update-rxvgg\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.551789 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn6jh\" (UniqueName: \"kubernetes.io/projected/9dbbc5fa-b903-4296-a3af-75524920938d-kube-api-access-rn6jh\") pod \"nova-cell1-94ac-account-create-update-rxvgg\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.564329 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.420564405 podStartE2EDuration="6.564309775s" podCreationTimestamp="2026-03-12 13:33:51 +0000 UTC" firstStartedPulling="2026-03-12 13:33:52.603779366 +0000 UTC m=+1451.052474762" lastFinishedPulling="2026-03-12 13:33:56.747524736 +0000 UTC m=+1455.196220132" observedRunningTime="2026-03-12 13:33:57.517081986 +0000 UTC m=+1455.965777382" watchObservedRunningTime="2026-03-12 13:33:57.564309775 +0000 UTC m=+1456.013005171" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.608627 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sckbb"] Mar 12 13:33:57 crc kubenswrapper[4778]: W0312 13:33:57.641936 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d11f6c3_3911_4a29_a65d_ef1f570d9b02.slice/crio-4880f510662a5f7132d196ce16de62950ef057f7d7af9da4c281a5ecfaedcca2 WatchSource:0}: Error finding container 4880f510662a5f7132d196ce16de62950ef057f7d7af9da4c281a5ecfaedcca2: Status 404 returned error can't find the container with id 4880f510662a5f7132d196ce16de62950ef057f7d7af9da4c281a5ecfaedcca2 Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.791862 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:33:57 crc kubenswrapper[4778]: I0312 13:33:57.931757 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x8nht"] Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.017416 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-dcf9-account-create-update-2rmjd"] Mar 12 13:33:58 crc kubenswrapper[4778]: W0312 13:33:58.032110 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ab681f_51c2_4723_b5b6_58c841185455.slice/crio-82b63dc5beaf8ac79964255d2c25b2ecfed64153aeb0366bfed3a3b034e6a33d WatchSource:0}: Error finding container 82b63dc5beaf8ac79964255d2c25b2ecfed64153aeb0366bfed3a3b034e6a33d: Status 404 returned error can't find the container with id 82b63dc5beaf8ac79964255d2c25b2ecfed64153aeb0366bfed3a3b034e6a33d Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.219399 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7d0f-account-create-update-t2rrl"] Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.245882 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2dh9w"] Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.285479 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-94ac-account-create-update-rxvgg"] Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.526463 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" event={"ID":"20ab681f-51c2-4723-b5b6-58c841185455","Type":"ContainerStarted","Data":"82b63dc5beaf8ac79964255d2c25b2ecfed64153aeb0366bfed3a3b034e6a33d"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.531042 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.531481 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-log" containerID="cri-o://0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260" gracePeriod=30 Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.531712 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-httpd" containerID="cri-o://7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f" gracePeriod=30 Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.542060 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" event={"ID":"9dbbc5fa-b903-4296-a3af-75524920938d","Type":"ContainerStarted","Data":"12e7d50aae58d04731b12f84ab30471cfbbda676ec2f653ddf3e5df4decd21b9"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.560673 4778 generic.go:334] "Generic (PLEG): container finished" podID="2d11f6c3-3911-4a29-a65d-ef1f570d9b02" containerID="bc734e634b97b1a5646716a6fc635d874255724a3ef890cee0802c7190db7d7c" exitCode=0 Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.561144 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sckbb" event={"ID":"2d11f6c3-3911-4a29-a65d-ef1f570d9b02","Type":"ContainerDied","Data":"bc734e634b97b1a5646716a6fc635d874255724a3ef890cee0802c7190db7d7c"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.561177 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sckbb" event={"ID":"2d11f6c3-3911-4a29-a65d-ef1f570d9b02","Type":"ContainerStarted","Data":"4880f510662a5f7132d196ce16de62950ef057f7d7af9da4c281a5ecfaedcca2"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.563475 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" event={"ID":"092c3556-0255-4e2f-b2c7-e22b8a3d8418","Type":"ContainerStarted","Data":"d33dd868a9cad44b520d800f2c0ee298247a48b4345e044c0c3ed20a58fc82c5"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.567485 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x8nht" event={"ID":"4ad4ff5d-b816-4bdd-97a7-8afd73afe583","Type":"ContainerStarted","Data":"e2a8d1e05ff7ff80a86b71f26e5fb5c7484878b8a9632420829088d85ad0fbaf"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.567531 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x8nht" event={"ID":"4ad4ff5d-b816-4bdd-97a7-8afd73afe583","Type":"ContainerStarted","Data":"d10b3450ecd206da6b0e8847141e39d6a2c9d193ad5e15dd86118dc2ded48a9f"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.568987 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2dh9w" event={"ID":"068c02bc-1daf-4029-84f9-39a395d5de3e","Type":"ContainerStarted","Data":"9ffe001ed68c4701eb0b1fec6393649f539478d01421e0e3619aa9a88a221722"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.578912 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fa757af-1c91-4b93-8916-5bbd99b8522e","Type":"ContainerStarted","Data":"91d9bb0428b135b688074df3a50b06e035e7a170ffbeb2f7181ae94e57fada1b"} Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.629740 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.629719716 podStartE2EDuration="4.629719716s" podCreationTimestamp="2026-03-12 13:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:58.624394405 +0000 UTC m=+1457.073089821" watchObservedRunningTime="2026-03-12 13:33:58.629719716 +0000 UTC m=+1457.078415122" Mar 12 13:33:58 crc kubenswrapper[4778]: I0312 13:33:58.644539 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-x8nht" podStartSLOduration=2.644517186 podStartE2EDuration="2.644517186s" podCreationTimestamp="2026-03-12 13:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:33:58.641199692 +0000 UTC m=+1457.089895088" watchObservedRunningTime="2026-03-12 13:33:58.644517186 +0000 UTC m=+1457.093212582" Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.588785 4778 generic.go:334] "Generic (PLEG): container finished" podID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerID="0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260" exitCode=143 Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.588995 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac92f5c5-e457-4915-a919-0dbe3df23ce8","Type":"ContainerDied","Data":"0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260"} Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.590568 4778 generic.go:334] "Generic (PLEG): container finished" podID="4ad4ff5d-b816-4bdd-97a7-8afd73afe583" containerID="e2a8d1e05ff7ff80a86b71f26e5fb5c7484878b8a9632420829088d85ad0fbaf" exitCode=0 Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.590638 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x8nht" event={"ID":"4ad4ff5d-b816-4bdd-97a7-8afd73afe583","Type":"ContainerDied","Data":"e2a8d1e05ff7ff80a86b71f26e5fb5c7484878b8a9632420829088d85ad0fbaf"} Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.594197 4778 generic.go:334] "Generic (PLEG): container finished" podID="068c02bc-1daf-4029-84f9-39a395d5de3e" containerID="37e7dd5198914cc6a22b8658dd88edbdbdabb2bfe43c9c4d07a686c73a997ca2" exitCode=0 Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.594261 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2dh9w" event={"ID":"068c02bc-1daf-4029-84f9-39a395d5de3e","Type":"ContainerDied","Data":"37e7dd5198914cc6a22b8658dd88edbdbdabb2bfe43c9c4d07a686c73a997ca2"} Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.596910 4778 generic.go:334] "Generic (PLEG): container finished" podID="20ab681f-51c2-4723-b5b6-58c841185455" containerID="be1aecf0f9c3a392b6320f1bb26caafd070dc71ba9db9be7a31ee5daf79e1a2d" exitCode=0 Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.596986 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" event={"ID":"20ab681f-51c2-4723-b5b6-58c841185455","Type":"ContainerDied","Data":"be1aecf0f9c3a392b6320f1bb26caafd070dc71ba9db9be7a31ee5daf79e1a2d"} Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.598629 4778 generic.go:334] "Generic (PLEG): container finished" podID="9dbbc5fa-b903-4296-a3af-75524920938d" containerID="099862fee239f9d58b6485a586d53c0613281de24cf1629f41917394af426901" exitCode=0 Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.598731 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" event={"ID":"9dbbc5fa-b903-4296-a3af-75524920938d","Type":"ContainerDied","Data":"099862fee239f9d58b6485a586d53c0613281de24cf1629f41917394af426901"} Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.602846 4778 generic.go:334] "Generic (PLEG): container finished" podID="092c3556-0255-4e2f-b2c7-e22b8a3d8418" containerID="2919ec7bf1dcd65b4aaec3b3c75478bba66c6d492f7b5c0064c9c993485c3e21" exitCode=0 Mar 12 13:33:59 crc kubenswrapper[4778]: I0312 13:33:59.603052 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" event={"ID":"092c3556-0255-4e2f-b2c7-e22b8a3d8418","Type":"ContainerDied","Data":"2919ec7bf1dcd65b4aaec3b3c75478bba66c6d492f7b5c0064c9c993485c3e21"} Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.112423 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sckbb" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.148813 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555374-lf8vj"] Mar 12 13:34:00 crc kubenswrapper[4778]: E0312 13:34:00.149390 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d11f6c3-3911-4a29-a65d-ef1f570d9b02" containerName="mariadb-database-create" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.149406 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d11f6c3-3911-4a29-a65d-ef1f570d9b02" containerName="mariadb-database-create" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.149595 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d11f6c3-3911-4a29-a65d-ef1f570d9b02" containerName="mariadb-database-create" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.150421 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.152819 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.155603 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.155833 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.168465 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555374-lf8vj"] Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.218217 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln84f\" (UniqueName: \"kubernetes.io/projected/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-kube-api-access-ln84f\") pod \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.218309 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-operator-scripts\") pod \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\" (UID: \"2d11f6c3-3911-4a29-a65d-ef1f570d9b02\") " Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.218682 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9h9r\" (UniqueName: \"kubernetes.io/projected/9d627011-802e-4075-9c56-43373d4c368e-kube-api-access-r9h9r\") pod \"auto-csr-approver-29555374-lf8vj\" (UID: \"9d627011-802e-4075-9c56-43373d4c368e\") " pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.219153 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d11f6c3-3911-4a29-a65d-ef1f570d9b02" (UID: "2d11f6c3-3911-4a29-a65d-ef1f570d9b02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.230571 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-kube-api-access-ln84f" (OuterVolumeSpecName: "kube-api-access-ln84f") pod "2d11f6c3-3911-4a29-a65d-ef1f570d9b02" (UID: "2d11f6c3-3911-4a29-a65d-ef1f570d9b02"). InnerVolumeSpecName "kube-api-access-ln84f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.326226 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9h9r\" (UniqueName: \"kubernetes.io/projected/9d627011-802e-4075-9c56-43373d4c368e-kube-api-access-r9h9r\") pod \"auto-csr-approver-29555374-lf8vj\" (UID: \"9d627011-802e-4075-9c56-43373d4c368e\") " pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.326645 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln84f\" (UniqueName: \"kubernetes.io/projected/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-kube-api-access-ln84f\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.326665 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d11f6c3-3911-4a29-a65d-ef1f570d9b02-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.349106 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9h9r\" (UniqueName: \"kubernetes.io/projected/9d627011-802e-4075-9c56-43373d4c368e-kube-api-access-r9h9r\") pod \"auto-csr-approver-29555374-lf8vj\" (UID: \"9d627011-802e-4075-9c56-43373d4c368e\") " pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.471737 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.615342 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sckbb" Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.616149 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sckbb" event={"ID":"2d11f6c3-3911-4a29-a65d-ef1f570d9b02","Type":"ContainerDied","Data":"4880f510662a5f7132d196ce16de62950ef057f7d7af9da4c281a5ecfaedcca2"} Mar 12 13:34:00 crc kubenswrapper[4778]: I0312 13:34:00.616173 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4880f510662a5f7132d196ce16de62950ef057f7d7af9da4c281a5ecfaedcca2" Mar 12 13:34:01 crc kubenswrapper[4778]: E0312 13:34:01.131683 4778 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.490158 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.505886 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.560129 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.560315 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.580491 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555374-lf8vj"] Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.584954 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.605899 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfdqt\" (UniqueName: \"kubernetes.io/projected/092c3556-0255-4e2f-b2c7-e22b8a3d8418-kube-api-access-nfdqt\") pod \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.606052 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn6jh\" (UniqueName: \"kubernetes.io/projected/9dbbc5fa-b903-4296-a3af-75524920938d-kube-api-access-rn6jh\") pod \"9dbbc5fa-b903-4296-a3af-75524920938d\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.607137 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092c3556-0255-4e2f-b2c7-e22b8a3d8418-operator-scripts\") pod \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\" (UID: \"092c3556-0255-4e2f-b2c7-e22b8a3d8418\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.607290 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbbc5fa-b903-4296-a3af-75524920938d-operator-scripts\") pod \"9dbbc5fa-b903-4296-a3af-75524920938d\" (UID: \"9dbbc5fa-b903-4296-a3af-75524920938d\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.614734 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/092c3556-0255-4e2f-b2c7-e22b8a3d8418-kube-api-access-nfdqt" (OuterVolumeSpecName: "kube-api-access-nfdqt") pod "092c3556-0255-4e2f-b2c7-e22b8a3d8418" (UID: "092c3556-0255-4e2f-b2c7-e22b8a3d8418"). InnerVolumeSpecName "kube-api-access-nfdqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.615335 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbbc5fa-b903-4296-a3af-75524920938d-kube-api-access-rn6jh" (OuterVolumeSpecName: "kube-api-access-rn6jh") pod "9dbbc5fa-b903-4296-a3af-75524920938d" (UID: "9dbbc5fa-b903-4296-a3af-75524920938d"). InnerVolumeSpecName "kube-api-access-rn6jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.615579 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dbbc5fa-b903-4296-a3af-75524920938d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9dbbc5fa-b903-4296-a3af-75524920938d" (UID: "9dbbc5fa-b903-4296-a3af-75524920938d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.615635 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/092c3556-0255-4e2f-b2c7-e22b8a3d8418-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "092c3556-0255-4e2f-b2c7-e22b8a3d8418" (UID: "092c3556-0255-4e2f-b2c7-e22b8a3d8418"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.654255 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2dh9w" event={"ID":"068c02bc-1daf-4029-84f9-39a395d5de3e","Type":"ContainerDied","Data":"9ffe001ed68c4701eb0b1fec6393649f539478d01421e0e3619aa9a88a221722"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.654558 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ffe001ed68c4701eb0b1fec6393649f539478d01421e0e3619aa9a88a221722" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.654705 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2dh9w" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.664456 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" event={"ID":"20ab681f-51c2-4723-b5b6-58c841185455","Type":"ContainerDied","Data":"82b63dc5beaf8ac79964255d2c25b2ecfed64153aeb0366bfed3a3b034e6a33d"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.664678 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82b63dc5beaf8ac79964255d2c25b2ecfed64153aeb0366bfed3a3b034e6a33d" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.664967 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dcf9-account-create-update-2rmjd" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.667537 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" event={"ID":"9dbbc5fa-b903-4296-a3af-75524920938d","Type":"ContainerDied","Data":"12e7d50aae58d04731b12f84ab30471cfbbda676ec2f653ddf3e5df4decd21b9"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.667649 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12e7d50aae58d04731b12f84ab30471cfbbda676ec2f653ddf3e5df4decd21b9" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.667795 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-94ac-account-create-update-rxvgg" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.672837 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" event={"ID":"092c3556-0255-4e2f-b2c7-e22b8a3d8418","Type":"ContainerDied","Data":"d33dd868a9cad44b520d800f2c0ee298247a48b4345e044c0c3ed20a58fc82c5"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.673248 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d33dd868a9cad44b520d800f2c0ee298247a48b4345e044c0c3ed20a58fc82c5" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.673360 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7d0f-account-create-update-t2rrl" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.681234 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" event={"ID":"9d627011-802e-4075-9c56-43373d4c368e","Type":"ContainerStarted","Data":"67e6d8b5b310afbd3ec49c3d001f4cd624f27618d4c040503f6dc47cb73ea130"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.690553 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x8nht" event={"ID":"4ad4ff5d-b816-4bdd-97a7-8afd73afe583","Type":"ContainerDied","Data":"d10b3450ecd206da6b0e8847141e39d6a2c9d193ad5e15dd86118dc2ded48a9f"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.690773 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d10b3450ecd206da6b0e8847141e39d6a2c9d193ad5e15dd86118dc2ded48a9f" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.690885 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x8nht" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.702339 4778 generic.go:334] "Generic (PLEG): container finished" podID="68092e68-04e5-4530-8d94-859789faeb94" containerID="8cdda802eadd8c68b3ba4b5b69b6a0fd021902af043f1083daaae42e4e3ba4bc" exitCode=0 Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.702454 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79ccdbbbbd-gl27l" event={"ID":"68092e68-04e5-4530-8d94-859789faeb94","Type":"ContainerDied","Data":"8cdda802eadd8c68b3ba4b5b69b6a0fd021902af043f1083daaae42e4e3ba4bc"} Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.708693 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klzfm\" (UniqueName: \"kubernetes.io/projected/068c02bc-1daf-4029-84f9-39a395d5de3e-kube-api-access-klzfm\") pod \"068c02bc-1daf-4029-84f9-39a395d5de3e\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.709025 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ab681f-51c2-4723-b5b6-58c841185455-operator-scripts\") pod \"20ab681f-51c2-4723-b5b6-58c841185455\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.709238 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvxrm\" (UniqueName: \"kubernetes.io/projected/20ab681f-51c2-4723-b5b6-58c841185455-kube-api-access-hvxrm\") pod \"20ab681f-51c2-4723-b5b6-58c841185455\" (UID: \"20ab681f-51c2-4723-b5b6-58c841185455\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.709345 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068c02bc-1daf-4029-84f9-39a395d5de3e-operator-scripts\") pod \"068c02bc-1daf-4029-84f9-39a395d5de3e\" (UID: \"068c02bc-1daf-4029-84f9-39a395d5de3e\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.709584 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-operator-scripts\") pod \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.709763 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwnk2\" (UniqueName: \"kubernetes.io/projected/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-kube-api-access-zwnk2\") pod \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\" (UID: \"4ad4ff5d-b816-4bdd-97a7-8afd73afe583\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.710438 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092c3556-0255-4e2f-b2c7-e22b8a3d8418-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.710783 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbbc5fa-b903-4296-a3af-75524920938d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.710880 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfdqt\" (UniqueName: \"kubernetes.io/projected/092c3556-0255-4e2f-b2c7-e22b8a3d8418-kube-api-access-nfdqt\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.710961 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn6jh\" (UniqueName: \"kubernetes.io/projected/9dbbc5fa-b903-4296-a3af-75524920938d-kube-api-access-rn6jh\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.713591 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/068c02bc-1daf-4029-84f9-39a395d5de3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "068c02bc-1daf-4029-84f9-39a395d5de3e" (UID: "068c02bc-1daf-4029-84f9-39a395d5de3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.713661 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20ab681f-51c2-4723-b5b6-58c841185455-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20ab681f-51c2-4723-b5b6-58c841185455" (UID: "20ab681f-51c2-4723-b5b6-58c841185455"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.713970 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ad4ff5d-b816-4bdd-97a7-8afd73afe583" (UID: "4ad4ff5d-b816-4bdd-97a7-8afd73afe583"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.716080 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-kube-api-access-zwnk2" (OuterVolumeSpecName: "kube-api-access-zwnk2") pod "4ad4ff5d-b816-4bdd-97a7-8afd73afe583" (UID: "4ad4ff5d-b816-4bdd-97a7-8afd73afe583"). InnerVolumeSpecName "kube-api-access-zwnk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.717040 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/068c02bc-1daf-4029-84f9-39a395d5de3e-kube-api-access-klzfm" (OuterVolumeSpecName: "kube-api-access-klzfm") pod "068c02bc-1daf-4029-84f9-39a395d5de3e" (UID: "068c02bc-1daf-4029-84f9-39a395d5de3e"). InnerVolumeSpecName "kube-api-access-klzfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.719285 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ab681f-51c2-4723-b5b6-58c841185455-kube-api-access-hvxrm" (OuterVolumeSpecName: "kube-api-access-hvxrm") pod "20ab681f-51c2-4723-b5b6-58c841185455" (UID: "20ab681f-51c2-4723-b5b6-58c841185455"). InnerVolumeSpecName "kube-api-access-hvxrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.722305 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.812477 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68092e68-04e5-4530-8d94-859789faeb94-logs\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.812833 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-public-tls-certs\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.812966 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-combined-ca-bundle\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.813817 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-config-data\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.814118 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnrxw\" (UniqueName: \"kubernetes.io/projected/68092e68-04e5-4530-8d94-859789faeb94-kube-api-access-jnrxw\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.813413 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68092e68-04e5-4530-8d94-859789faeb94-logs" (OuterVolumeSpecName: "logs") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.814482 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-internal-tls-certs\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.814989 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-scripts\") pod \"68092e68-04e5-4530-8d94-859789faeb94\" (UID: \"68092e68-04e5-4530-8d94-859789faeb94\") " Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.816511 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.816922 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwnk2\" (UniqueName: \"kubernetes.io/projected/4ad4ff5d-b816-4bdd-97a7-8afd73afe583-kube-api-access-zwnk2\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.817043 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klzfm\" (UniqueName: \"kubernetes.io/projected/068c02bc-1daf-4029-84f9-39a395d5de3e-kube-api-access-klzfm\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.817361 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68092e68-04e5-4530-8d94-859789faeb94-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.817825 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ab681f-51c2-4723-b5b6-58c841185455-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.817933 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvxrm\" (UniqueName: \"kubernetes.io/projected/20ab681f-51c2-4723-b5b6-58c841185455-kube-api-access-hvxrm\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.818241 4778 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068c02bc-1daf-4029-84f9-39a395d5de3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.819813 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-scripts" (OuterVolumeSpecName: "scripts") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.820985 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68092e68-04e5-4530-8d94-859789faeb94-kube-api-access-jnrxw" (OuterVolumeSpecName: "kube-api-access-jnrxw") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "kube-api-access-jnrxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.902739 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-config-data" (OuterVolumeSpecName: "config-data") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.907414 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.920148 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.920208 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.920222 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnrxw\" (UniqueName: \"kubernetes.io/projected/68092e68-04e5-4530-8d94-859789faeb94-kube-api-access-jnrxw\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.920236 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.944982 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:01 crc kubenswrapper[4778]: I0312 13:34:01.966261 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "68092e68-04e5-4530-8d94-859789faeb94" (UID: "68092e68-04e5-4530-8d94-859789faeb94"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.022566 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.022645 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68092e68-04e5-4530-8d94-859789faeb94-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.107714 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.227736 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.227804 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-logs\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.227935 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-public-tls-certs\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.227960 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-httpd-run\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.227977 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pklz\" (UniqueName: \"kubernetes.io/projected/ac92f5c5-e457-4915-a919-0dbe3df23ce8-kube-api-access-5pklz\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.228035 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-combined-ca-bundle\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.228125 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-config-data\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.228171 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-scripts\") pod \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\" (UID: \"ac92f5c5-e457-4915-a919-0dbe3df23ce8\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.230891 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.233030 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.233621 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-logs" (OuterVolumeSpecName: "logs") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.236444 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.251427 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-scripts" (OuterVolumeSpecName: "scripts") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.259044 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac92f5c5-e457-4915-a919-0dbe3df23ce8-kube-api-access-5pklz" (OuterVolumeSpecName: "kube-api-access-5pklz") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "kube-api-access-5pklz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.286093 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.324471 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.334719 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.334778 4778 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.334795 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac92f5c5-e457-4915-a919-0dbe3df23ce8-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.334806 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.335010 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pklz\" (UniqueName: \"kubernetes.io/projected/ac92f5c5-e457-4915-a919-0dbe3df23ce8-kube-api-access-5pklz\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.335023 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.352442 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-config-data" (OuterVolumeSpecName: "config-data") pod "ac92f5c5-e457-4915-a919-0dbe3df23ce8" (UID: "ac92f5c5-e457-4915-a919-0dbe3df23ce8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.359931 4778 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.436440 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac92f5c5-e457-4915-a919-0dbe3df23ce8-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.436473 4778 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.716132 4778 generic.go:334] "Generic (PLEG): container finished" podID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerID="7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f" exitCode=0 Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.716213 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac92f5c5-e457-4915-a919-0dbe3df23ce8","Type":"ContainerDied","Data":"7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f"} Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.716244 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac92f5c5-e457-4915-a919-0dbe3df23ce8","Type":"ContainerDied","Data":"de7bb235534c3c0c1a6530e35fd6d03d222f02129ca88b49fda3a8c136ab05b7"} Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.716263 4778 scope.go:117] "RemoveContainer" containerID="7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.716401 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.722845 4778 generic.go:334] "Generic (PLEG): container finished" podID="e34be903-da25-4cdb-9298-2d53fdce0276" containerID="76d710be6da7b239e82f6228977b9799ccd95f2824b23913a0585897e926dd74" exitCode=0 Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.722923 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-678c76989b-8x56d" event={"ID":"e34be903-da25-4cdb-9298-2d53fdce0276","Type":"ContainerDied","Data":"76d710be6da7b239e82f6228977b9799ccd95f2824b23913a0585897e926dd74"} Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.730915 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79ccdbbbbd-gl27l" event={"ID":"68092e68-04e5-4530-8d94-859789faeb94","Type":"ContainerDied","Data":"6225b0b7ab31929807b7000d1c797565cb38b8453f9487cc91d0a8fcf517ace6"} Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.731023 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79ccdbbbbd-gl27l" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.804836 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.823771 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.827777 4778 scope.go:117] "RemoveContainer" containerID="0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.846718 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.869486 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-79ccdbbbbd-gl27l"] Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.881312 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883282 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-httpd" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883316 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-httpd" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883356 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ab681f-51c2-4723-b5b6-58c841185455" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883369 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ab681f-51c2-4723-b5b6-58c841185455" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883393 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-log" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883406 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-log" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883442 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092c3556-0255-4e2f-b2c7-e22b8a3d8418" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883454 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="092c3556-0255-4e2f-b2c7-e22b8a3d8418" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883468 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-api" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883480 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-api" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883504 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbbc5fa-b903-4296-a3af-75524920938d" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883512 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbbc5fa-b903-4296-a3af-75524920938d" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883522 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068c02bc-1daf-4029-84f9-39a395d5de3e" containerName="mariadb-database-create" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883531 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="068c02bc-1daf-4029-84f9-39a395d5de3e" containerName="mariadb-database-create" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883562 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-log" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883575 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-log" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883602 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-httpd" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883615 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-httpd" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883645 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad4ff5d-b816-4bdd-97a7-8afd73afe583" containerName="mariadb-database-create" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883654 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad4ff5d-b816-4bdd-97a7-8afd73afe583" containerName="mariadb-database-create" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.883667 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-api" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.883675 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-api" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885383 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="092c3556-0255-4e2f-b2c7-e22b8a3d8418" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885410 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-log" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885420 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-api" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885431 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad4ff5d-b816-4bdd-97a7-8afd73afe583" containerName="mariadb-database-create" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885456 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbbc5fa-b903-4296-a3af-75524920938d" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885467 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-httpd" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885488 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="068c02bc-1daf-4029-84f9-39a395d5de3e" containerName="mariadb-database-create" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885508 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ab681f-51c2-4723-b5b6-58c841185455" containerName="mariadb-account-create-update" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885530 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" containerName="glance-httpd" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885544 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" containerName="neutron-api" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.885560 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="68092e68-04e5-4530-8d94-859789faeb94" containerName="placement-log" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.902931 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.910626 4778 scope.go:117] "RemoveContainer" containerID="7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.911097 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.911268 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.911540 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f\": container with ID starting with 7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f not found: ID does not exist" containerID="7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.913232 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f"} err="failed to get container status \"7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f\": rpc error: code = NotFound desc = could not find container \"7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f\": container with ID starting with 7ffa53ad28bc0a9b2bc486f510ba4d00037b24ab1288b231f7af2e96baabc19f not found: ID does not exist" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.913305 4778 scope.go:117] "RemoveContainer" containerID="0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.911602 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-79ccdbbbbd-gl27l"] Mar 12 13:34:02 crc kubenswrapper[4778]: E0312 13:34:02.924334 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260\": container with ID starting with 0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260 not found: ID does not exist" containerID="0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.924426 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260"} err="failed to get container status \"0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260\": rpc error: code = NotFound desc = could not find container \"0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260\": container with ID starting with 0aad8b26d122f41726307150aa529e2cdb5f197081e19ece7255f5b8aa07d260 not found: ID does not exist" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.924472 4778 scope.go:117] "RemoveContainer" containerID="8cdda802eadd8c68b3ba4b5b69b6a0fd021902af043f1083daaae42e4e3ba4bc" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.941117 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.950456 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-combined-ca-bundle\") pod \"e34be903-da25-4cdb-9298-2d53fdce0276\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.950570 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-httpd-config\") pod \"e34be903-da25-4cdb-9298-2d53fdce0276\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.950649 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-config\") pod \"e34be903-da25-4cdb-9298-2d53fdce0276\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.950734 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-ovndb-tls-certs\") pod \"e34be903-da25-4cdb-9298-2d53fdce0276\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.950787 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgmt8\" (UniqueName: \"kubernetes.io/projected/e34be903-da25-4cdb-9298-2d53fdce0276-kube-api-access-cgmt8\") pod \"e34be903-da25-4cdb-9298-2d53fdce0276\" (UID: \"e34be903-da25-4cdb-9298-2d53fdce0276\") " Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.957617 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e34be903-da25-4cdb-9298-2d53fdce0276" (UID: "e34be903-da25-4cdb-9298-2d53fdce0276"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.958301 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34be903-da25-4cdb-9298-2d53fdce0276-kube-api-access-cgmt8" (OuterVolumeSpecName: "kube-api-access-cgmt8") pod "e34be903-da25-4cdb-9298-2d53fdce0276" (UID: "e34be903-da25-4cdb-9298-2d53fdce0276"). InnerVolumeSpecName "kube-api-access-cgmt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:02 crc kubenswrapper[4778]: I0312 13:34:02.971538 4778 scope.go:117] "RemoveContainer" containerID="be846a255557e511860dc7bc1b884d65bc6e48bfb1b98ae1316cb74617623c2b" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.021109 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e34be903-da25-4cdb-9298-2d53fdce0276" (UID: "e34be903-da25-4cdb-9298-2d53fdce0276"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.025739 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-config" (OuterVolumeSpecName: "config") pod "e34be903-da25-4cdb-9298-2d53fdce0276" (UID: "e34be903-da25-4cdb-9298-2d53fdce0276"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.040722 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e34be903-da25-4cdb-9298-2d53fdce0276" (UID: "e34be903-da25-4cdb-9298-2d53fdce0276"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057233 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tjbq\" (UniqueName: \"kubernetes.io/projected/81c1a05c-5642-43d4-8a7b-229330168332-kube-api-access-9tjbq\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057305 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057370 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057461 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-scripts\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057522 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81c1a05c-5642-43d4-8a7b-229330168332-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057543 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81c1a05c-5642-43d4-8a7b-229330168332-logs\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057563 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-config-data\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057587 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057639 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057662 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057672 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057683 4778 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e34be903-da25-4cdb-9298-2d53fdce0276-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.057692 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgmt8\" (UniqueName: \"kubernetes.io/projected/e34be903-da25-4cdb-9298-2d53fdce0276-kube-api-access-cgmt8\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.158859 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-scripts\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.158925 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81c1a05c-5642-43d4-8a7b-229330168332-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.158954 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81c1a05c-5642-43d4-8a7b-229330168332-logs\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.158977 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-config-data\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.159009 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.159038 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tjbq\" (UniqueName: \"kubernetes.io/projected/81c1a05c-5642-43d4-8a7b-229330168332-kube-api-access-9tjbq\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.159086 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.159150 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.160325 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81c1a05c-5642-43d4-8a7b-229330168332-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.160364 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81c1a05c-5642-43d4-8a7b-229330168332-logs\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.160916 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") device mount path \"/mnt/openstack/pv17\"" pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.164146 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.172730 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-config-data\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.177527 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.178204 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81c1a05c-5642-43d4-8a7b-229330168332-scripts\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.180354 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tjbq\" (UniqueName: \"kubernetes.io/projected/81c1a05c-5642-43d4-8a7b-229330168332-kube-api-access-9tjbq\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.202915 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"81c1a05c-5642-43d4-8a7b-229330168332\") " pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.233540 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.740730 4778 generic.go:334] "Generic (PLEG): container finished" podID="9d627011-802e-4075-9c56-43373d4c368e" containerID="d817d5a09b7856e71332e283d84fe3ea296ae040cb7e986cd73c433864a99c34" exitCode=0 Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.741162 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" event={"ID":"9d627011-802e-4075-9c56-43373d4c368e","Type":"ContainerDied","Data":"d817d5a09b7856e71332e283d84fe3ea296ae040cb7e986cd73c433864a99c34"} Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.744904 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-678c76989b-8x56d" event={"ID":"e34be903-da25-4cdb-9298-2d53fdce0276","Type":"ContainerDied","Data":"73ff3b874391ffdc31812d5d85f13741c2920b13dddb21f9bdace835187b0822"} Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.744941 4778 scope.go:117] "RemoveContainer" containerID="7423051fcfb7c12e56b049e90be94c641f82520ceab5181c7fcca6713588c77f" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.745048 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-678c76989b-8x56d" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.777990 4778 scope.go:117] "RemoveContainer" containerID="76d710be6da7b239e82f6228977b9799ccd95f2824b23913a0585897e926dd74" Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.809730 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.818407 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-678c76989b-8x56d"] Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.826948 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.827297 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-central-agent" containerID="cri-o://e925c9c4c7aa08744211c517c124058aade623d45fd2e02df2777b4f2df794b2" gracePeriod=30 Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.827678 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="proxy-httpd" containerID="cri-o://013c8995ca90639ba33078e18954ed308111d321639179db05cd00d19ef56702" gracePeriod=30 Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.827726 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="sg-core" containerID="cri-o://86ce4b4705307dac2aa00c6fc4314d927b33960ed30dc0799f79715a9adfcdf9" gracePeriod=30 Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.827761 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-notification-agent" containerID="cri-o://8581110fc2e8206867368b8c4ae7af28cb79d5341dbf6b92ea91def7d2e28eb6" gracePeriod=30 Mar 12 13:34:03 crc kubenswrapper[4778]: I0312 13:34:03.834089 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-678c76989b-8x56d"] Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.285616 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68092e68-04e5-4530-8d94-859789faeb94" path="/var/lib/kubelet/pods/68092e68-04e5-4530-8d94-859789faeb94/volumes" Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.286561 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac92f5c5-e457-4915-a919-0dbe3df23ce8" path="/var/lib/kubelet/pods/ac92f5c5-e457-4915-a919-0dbe3df23ce8/volumes" Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.287176 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e34be903-da25-4cdb-9298-2d53fdce0276" path="/var/lib/kubelet/pods/e34be903-da25-4cdb-9298-2d53fdce0276/volumes" Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.763315 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"81c1a05c-5642-43d4-8a7b-229330168332","Type":"ContainerStarted","Data":"84f7024aeceefe982e2cee4c0fa46027923ecf1b4ef431d7fe2e34b1d5e3f2e6"} Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.763375 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"81c1a05c-5642-43d4-8a7b-229330168332","Type":"ContainerStarted","Data":"f128424220ef8224b8053627c5d48ddef6508a615314fe7e2b95951288127305"} Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.767464 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerID="013c8995ca90639ba33078e18954ed308111d321639179db05cd00d19ef56702" exitCode=0 Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.767492 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerID="86ce4b4705307dac2aa00c6fc4314d927b33960ed30dc0799f79715a9adfcdf9" exitCode=2 Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.767500 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerID="8581110fc2e8206867368b8c4ae7af28cb79d5341dbf6b92ea91def7d2e28eb6" exitCode=0 Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.767569 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerDied","Data":"013c8995ca90639ba33078e18954ed308111d321639179db05cd00d19ef56702"} Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.767671 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerDied","Data":"86ce4b4705307dac2aa00c6fc4314d927b33960ed30dc0799f79715a9adfcdf9"} Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.767692 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerDied","Data":"8581110fc2e8206867368b8c4ae7af28cb79d5341dbf6b92ea91def7d2e28eb6"} Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.773124 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.773165 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.820682 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:04 crc kubenswrapper[4778]: I0312 13:34:04.872382 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.090991 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.209644 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9h9r\" (UniqueName: \"kubernetes.io/projected/9d627011-802e-4075-9c56-43373d4c368e-kube-api-access-r9h9r\") pod \"9d627011-802e-4075-9c56-43373d4c368e\" (UID: \"9d627011-802e-4075-9c56-43373d4c368e\") " Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.269550 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d627011-802e-4075-9c56-43373d4c368e-kube-api-access-r9h9r" (OuterVolumeSpecName: "kube-api-access-r9h9r") pod "9d627011-802e-4075-9c56-43373d4c368e" (UID: "9d627011-802e-4075-9c56-43373d4c368e"). InnerVolumeSpecName "kube-api-access-r9h9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.312483 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9h9r\" (UniqueName: \"kubernetes.io/projected/9d627011-802e-4075-9c56-43373d4c368e-kube-api-access-r9h9r\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.777088 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" event={"ID":"9d627011-802e-4075-9c56-43373d4c368e","Type":"ContainerDied","Data":"67e6d8b5b310afbd3ec49c3d001f4cd624f27618d4c040503f6dc47cb73ea130"} Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.777179 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e6d8b5b310afbd3ec49c3d001f4cd624f27618d4c040503f6dc47cb73ea130" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.777141 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555374-lf8vj" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.784215 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"81c1a05c-5642-43d4-8a7b-229330168332","Type":"ContainerStarted","Data":"c5c75538574d506f02760e3d1dc542bf94ef25dccf27637b10ea3a8078431e9e"} Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.784676 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:05 crc kubenswrapper[4778]: I0312 13:34:05.784725 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:06 crc kubenswrapper[4778]: I0312 13:34:06.144169 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.144152669 podStartE2EDuration="4.144152669s" podCreationTimestamp="2026-03-12 13:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:05.808982752 +0000 UTC m=+1464.257678148" watchObservedRunningTime="2026-03-12 13:34:06.144152669 +0000 UTC m=+1464.592848065" Mar 12 13:34:06 crc kubenswrapper[4778]: I0312 13:34:06.180557 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555368-d2cpg"] Mar 12 13:34:06 crc kubenswrapper[4778]: I0312 13:34:06.188574 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555368-d2cpg"] Mar 12 13:34:06 crc kubenswrapper[4778]: I0312 13:34:06.266106 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20d587ee-7b57-4b99-a800-c6d46322d799" path="/var/lib/kubelet/pods/20d587ee-7b57-4b99-a800-c6d46322d799/volumes" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.414650 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6b6mv"] Mar 12 13:34:07 crc kubenswrapper[4778]: E0312 13:34:07.415389 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d627011-802e-4075-9c56-43373d4c368e" containerName="oc" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.415402 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d627011-802e-4075-9c56-43373d4c368e" containerName="oc" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.415580 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d627011-802e-4075-9c56-43373d4c368e" containerName="oc" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.416292 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.418333 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.418737 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.418742 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bjjj5" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.431740 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6b6mv"] Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.446512 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q225s\" (UniqueName: \"kubernetes.io/projected/fe24691f-9019-44ec-85bf-b477c53f05ec-kube-api-access-q225s\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.446852 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-config-data\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.447138 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-scripts\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.447245 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.548705 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-scripts\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.549730 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.549859 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q225s\" (UniqueName: \"kubernetes.io/projected/fe24691f-9019-44ec-85bf-b477c53f05ec-kube-api-access-q225s\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.549943 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-config-data\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.555725 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-config-data\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.555738 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-scripts\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.556736 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.571687 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q225s\" (UniqueName: \"kubernetes.io/projected/fe24691f-9019-44ec-85bf-b477c53f05ec-kube-api-access-q225s\") pod \"nova-cell0-conductor-db-sync-6b6mv\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.733716 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.823975 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerID="e925c9c4c7aa08744211c517c124058aade623d45fd2e02df2777b4f2df794b2" exitCode=0 Mar 12 13:34:07 crc kubenswrapper[4778]: I0312 13:34:07.824027 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerDied","Data":"e925c9c4c7aa08744211c517c124058aade623d45fd2e02df2777b4f2df794b2"} Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.144490 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.160765 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-log-httpd\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.160874 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-combined-ca-bundle\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.160938 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-sg-core-conf-yaml\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161004 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrpfb\" (UniqueName: \"kubernetes.io/projected/0bf8c182-c9d5-4011-b28c-c4f557a8071c-kube-api-access-nrpfb\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161080 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-config-data\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161157 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-scripts\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161324 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-run-httpd\") pod \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\" (UID: \"0bf8c182-c9d5-4011-b28c-c4f557a8071c\") " Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161337 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161859 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.161881 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.168859 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-scripts" (OuterVolumeSpecName: "scripts") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.180919 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf8c182-c9d5-4011-b28c-c4f557a8071c-kube-api-access-nrpfb" (OuterVolumeSpecName: "kube-api-access-nrpfb") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "kube-api-access-nrpfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.204481 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.256586 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.264375 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bf8c182-c9d5-4011-b28c-c4f557a8071c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.264401 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.264411 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.264437 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrpfb\" (UniqueName: \"kubernetes.io/projected/0bf8c182-c9d5-4011-b28c-c4f557a8071c-kube-api-access-nrpfb\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.264447 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.300091 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-config-data" (OuterVolumeSpecName: "config-data") pod "0bf8c182-c9d5-4011-b28c-c4f557a8071c" (UID: "0bf8c182-c9d5-4011-b28c-c4f557a8071c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:08 crc kubenswrapper[4778]: W0312 13:34:08.304567 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe24691f_9019_44ec_85bf_b477c53f05ec.slice/crio-28ffeba46951880404aa9c0e4e9f8643f9909fdcbeeecb541b1919d958482b53 WatchSource:0}: Error finding container 28ffeba46951880404aa9c0e4e9f8643f9909fdcbeeecb541b1919d958482b53: Status 404 returned error can't find the container with id 28ffeba46951880404aa9c0e4e9f8643f9909fdcbeeecb541b1919d958482b53 Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.366145 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf8c182-c9d5-4011-b28c-c4f557a8071c-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.380499 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6b6mv"] Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.380590 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.380659 4778 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.389552 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.839609 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bf8c182-c9d5-4011-b28c-c4f557a8071c","Type":"ContainerDied","Data":"0f3aa121caf2c1a6a7f5f32c4c791af4c518cf20357d26f2062f2e017c408468"} Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.839697 4778 scope.go:117] "RemoveContainer" containerID="013c8995ca90639ba33078e18954ed308111d321639179db05cd00d19ef56702" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.840741 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.849949 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" event={"ID":"fe24691f-9019-44ec-85bf-b477c53f05ec","Type":"ContainerStarted","Data":"28ffeba46951880404aa9c0e4e9f8643f9909fdcbeeecb541b1919d958482b53"} Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.864540 4778 scope.go:117] "RemoveContainer" containerID="86ce4b4705307dac2aa00c6fc4314d927b33960ed30dc0799f79715a9adfcdf9" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.891714 4778 scope.go:117] "RemoveContainer" containerID="8581110fc2e8206867368b8c4ae7af28cb79d5341dbf6b92ea91def7d2e28eb6" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.902301 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.915013 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.929905 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930098 4778 scope.go:117] "RemoveContainer" containerID="e925c9c4c7aa08744211c517c124058aade623d45fd2e02df2777b4f2df794b2" Mar 12 13:34:08 crc kubenswrapper[4778]: E0312 13:34:08.930414 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="proxy-httpd" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930436 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="proxy-httpd" Mar 12 13:34:08 crc kubenswrapper[4778]: E0312 13:34:08.930466 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-notification-agent" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930476 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-notification-agent" Mar 12 13:34:08 crc kubenswrapper[4778]: E0312 13:34:08.930496 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="sg-core" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930504 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="sg-core" Mar 12 13:34:08 crc kubenswrapper[4778]: E0312 13:34:08.930527 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-central-agent" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930535 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-central-agent" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930741 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="proxy-httpd" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930757 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-notification-agent" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930768 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="ceilometer-central-agent" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.930790 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" containerName="sg-core" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.933443 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.936021 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.936180 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:34:08 crc kubenswrapper[4778]: I0312 13:34:08.942786 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.080663 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-scripts\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.080718 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-log-httpd\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.080750 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.080937 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft9qb\" (UniqueName: \"kubernetes.io/projected/f7a2b0f7-9321-4f29-aa01-0acbc528f757-kube-api-access-ft9qb\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.081013 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-config-data\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.081060 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-run-httpd\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.081112 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.183065 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-config-data\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.183134 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-run-httpd\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.183178 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.183774 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-run-httpd\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.183864 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-scripts\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.184355 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-log-httpd\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.184891 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-log-httpd\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.184981 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.187243 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft9qb\" (UniqueName: \"kubernetes.io/projected/f7a2b0f7-9321-4f29-aa01-0acbc528f757-kube-api-access-ft9qb\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.191231 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-config-data\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.191876 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.192577 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.193005 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-scripts\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.207827 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft9qb\" (UniqueName: \"kubernetes.io/projected/f7a2b0f7-9321-4f29-aa01-0acbc528f757-kube-api-access-ft9qb\") pod \"ceilometer-0\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.256922 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.708628 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:09 crc kubenswrapper[4778]: I0312 13:34:09.862365 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerStarted","Data":"b0892c0b95fcd75a23c8ced61d0a214155429e8763fde0aacd038b1f9445ab5d"} Mar 12 13:34:10 crc kubenswrapper[4778]: I0312 13:34:10.266167 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf8c182-c9d5-4011-b28c-c4f557a8071c" path="/var/lib/kubelet/pods/0bf8c182-c9d5-4011-b28c-c4f557a8071c/volumes" Mar 12 13:34:10 crc kubenswrapper[4778]: I0312 13:34:10.877669 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerStarted","Data":"07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67"} Mar 12 13:34:11 crc kubenswrapper[4778]: I0312 13:34:11.888697 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerStarted","Data":"5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d"} Mar 12 13:34:13 crc kubenswrapper[4778]: I0312 13:34:13.235049 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 13:34:13 crc kubenswrapper[4778]: I0312 13:34:13.236625 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 13:34:13 crc kubenswrapper[4778]: I0312 13:34:13.281577 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 13:34:13 crc kubenswrapper[4778]: I0312 13:34:13.306545 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 13:34:13 crc kubenswrapper[4778]: I0312 13:34:13.909150 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 13:34:13 crc kubenswrapper[4778]: I0312 13:34:13.909210 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 13:34:15 crc kubenswrapper[4778]: I0312 13:34:15.845264 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 13:34:15 crc kubenswrapper[4778]: I0312 13:34:15.847854 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 13:34:16 crc kubenswrapper[4778]: I0312 13:34:16.938114 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerStarted","Data":"90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c"} Mar 12 13:34:16 crc kubenswrapper[4778]: I0312 13:34:16.941162 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" event={"ID":"fe24691f-9019-44ec-85bf-b477c53f05ec","Type":"ContainerStarted","Data":"e3b15e2b52f4e1dd648d1cbcdd4c757ead8e48ae1ed5c998744e64dfa8993e67"} Mar 12 13:34:16 crc kubenswrapper[4778]: I0312 13:34:16.971247 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" podStartSLOduration=2.26765603 podStartE2EDuration="9.971226618s" podCreationTimestamp="2026-03-12 13:34:07 +0000 UTC" firstStartedPulling="2026-03-12 13:34:08.308093222 +0000 UTC m=+1466.756788618" lastFinishedPulling="2026-03-12 13:34:16.01166381 +0000 UTC m=+1474.460359206" observedRunningTime="2026-03-12 13:34:16.969641053 +0000 UTC m=+1475.418336449" watchObservedRunningTime="2026-03-12 13:34:16.971226618 +0000 UTC m=+1475.419922014" Mar 12 13:34:18 crc kubenswrapper[4778]: I0312 13:34:18.962488 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerStarted","Data":"8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf"} Mar 12 13:34:18 crc kubenswrapper[4778]: I0312 13:34:18.962930 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:34:18 crc kubenswrapper[4778]: I0312 13:34:18.995326 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.206344647 podStartE2EDuration="10.994162761s" podCreationTimestamp="2026-03-12 13:34:08 +0000 UTC" firstStartedPulling="2026-03-12 13:34:09.726357941 +0000 UTC m=+1468.175053337" lastFinishedPulling="2026-03-12 13:34:18.514176055 +0000 UTC m=+1476.962871451" observedRunningTime="2026-03-12 13:34:18.983852428 +0000 UTC m=+1477.432547824" watchObservedRunningTime="2026-03-12 13:34:18.994162761 +0000 UTC m=+1477.442858157" Mar 12 13:34:29 crc kubenswrapper[4778]: I0312 13:34:29.070671 4778 generic.go:334] "Generic (PLEG): container finished" podID="fe24691f-9019-44ec-85bf-b477c53f05ec" containerID="e3b15e2b52f4e1dd648d1cbcdd4c757ead8e48ae1ed5c998744e64dfa8993e67" exitCode=0 Mar 12 13:34:29 crc kubenswrapper[4778]: I0312 13:34:29.071438 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" event={"ID":"fe24691f-9019-44ec-85bf-b477c53f05ec","Type":"ContainerDied","Data":"e3b15e2b52f4e1dd648d1cbcdd4c757ead8e48ae1ed5c998744e64dfa8993e67"} Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.416036 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.526751 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q225s\" (UniqueName: \"kubernetes.io/projected/fe24691f-9019-44ec-85bf-b477c53f05ec-kube-api-access-q225s\") pod \"fe24691f-9019-44ec-85bf-b477c53f05ec\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.526952 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-combined-ca-bundle\") pod \"fe24691f-9019-44ec-85bf-b477c53f05ec\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.527009 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-scripts\") pod \"fe24691f-9019-44ec-85bf-b477c53f05ec\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.527096 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-config-data\") pod \"fe24691f-9019-44ec-85bf-b477c53f05ec\" (UID: \"fe24691f-9019-44ec-85bf-b477c53f05ec\") " Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.536946 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-scripts" (OuterVolumeSpecName: "scripts") pod "fe24691f-9019-44ec-85bf-b477c53f05ec" (UID: "fe24691f-9019-44ec-85bf-b477c53f05ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.537061 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe24691f-9019-44ec-85bf-b477c53f05ec-kube-api-access-q225s" (OuterVolumeSpecName: "kube-api-access-q225s") pod "fe24691f-9019-44ec-85bf-b477c53f05ec" (UID: "fe24691f-9019-44ec-85bf-b477c53f05ec"). InnerVolumeSpecName "kube-api-access-q225s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.554750 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe24691f-9019-44ec-85bf-b477c53f05ec" (UID: "fe24691f-9019-44ec-85bf-b477c53f05ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.555924 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-config-data" (OuterVolumeSpecName: "config-data") pod "fe24691f-9019-44ec-85bf-b477c53f05ec" (UID: "fe24691f-9019-44ec-85bf-b477c53f05ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.632785 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.632835 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q225s\" (UniqueName: \"kubernetes.io/projected/fe24691f-9019-44ec-85bf-b477c53f05ec-kube-api-access-q225s\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.632851 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:30 crc kubenswrapper[4778]: I0312 13:34:30.632862 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe24691f-9019-44ec-85bf-b477c53f05ec-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.091580 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" event={"ID":"fe24691f-9019-44ec-85bf-b477c53f05ec","Type":"ContainerDied","Data":"28ffeba46951880404aa9c0e4e9f8643f9909fdcbeeecb541b1919d958482b53"} Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.091627 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28ffeba46951880404aa9c0e4e9f8643f9909fdcbeeecb541b1919d958482b53" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.091712 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6b6mv" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.195598 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:34:31 crc kubenswrapper[4778]: E0312 13:34:31.196157 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe24691f-9019-44ec-85bf-b477c53f05ec" containerName="nova-cell0-conductor-db-sync" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.196204 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe24691f-9019-44ec-85bf-b477c53f05ec" containerName="nova-cell0-conductor-db-sync" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.196541 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe24691f-9019-44ec-85bf-b477c53f05ec" containerName="nova-cell0-conductor-db-sync" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.197385 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.200066 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bjjj5" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.200326 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.207936 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.349448 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.351785 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.352879 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xht95\" (UniqueName: \"kubernetes.io/projected/7733a48b-2bc4-4372-a222-37bb8ea04b6d-kube-api-access-xht95\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.457797 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xht95\" (UniqueName: \"kubernetes.io/projected/7733a48b-2bc4-4372-a222-37bb8ea04b6d-kube-api-access-xht95\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.457876 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.458006 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.469475 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.469832 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.474828 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xht95\" (UniqueName: \"kubernetes.io/projected/7733a48b-2bc4-4372-a222-37bb8ea04b6d-kube-api-access-xht95\") pod \"nova-cell0-conductor-0\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:31 crc kubenswrapper[4778]: I0312 13:34:31.516165 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:32 crc kubenswrapper[4778]: I0312 13:34:32.021475 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:34:32 crc kubenswrapper[4778]: I0312 13:34:32.101327 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7733a48b-2bc4-4372-a222-37bb8ea04b6d","Type":"ContainerStarted","Data":"be21932167b4499354351c16537055f7625655a6cd039664d1dd7fad790b8909"} Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.120039 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7733a48b-2bc4-4372-a222-37bb8ea04b6d","Type":"ContainerStarted","Data":"7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731"} Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.120518 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.151384 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.151361701 podStartE2EDuration="2.151361701s" podCreationTimestamp="2026-03-12 13:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:33.144750353 +0000 UTC m=+1491.593445769" watchObservedRunningTime="2026-03-12 13:34:33.151361701 +0000 UTC m=+1491.600057097" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.312532 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jfzqk"] Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.314444 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.323591 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jfzqk"] Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.401723 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-utilities\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.401786 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-catalog-content\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.402340 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frrb7\" (UniqueName: \"kubernetes.io/projected/1b7d48c4-04cd-481a-976d-19e57a28a1d9-kube-api-access-frrb7\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.505147 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-utilities\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.505212 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-catalog-content\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.505297 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frrb7\" (UniqueName: \"kubernetes.io/projected/1b7d48c4-04cd-481a-976d-19e57a28a1d9-kube-api-access-frrb7\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.505803 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-utilities\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.505832 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-catalog-content\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.525340 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frrb7\" (UniqueName: \"kubernetes.io/projected/1b7d48c4-04cd-481a-976d-19e57a28a1d9-kube-api-access-frrb7\") pod \"redhat-operators-jfzqk\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:33 crc kubenswrapper[4778]: I0312 13:34:33.642005 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:34 crc kubenswrapper[4778]: I0312 13:34:34.156016 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jfzqk"] Mar 12 13:34:34 crc kubenswrapper[4778]: W0312 13:34:34.161855 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b7d48c4_04cd_481a_976d_19e57a28a1d9.slice/crio-68397e437c2fb3791ad659ab6abc466e5cb77e5b97a5ba4bc1bb524e525fb6c3 WatchSource:0}: Error finding container 68397e437c2fb3791ad659ab6abc466e5cb77e5b97a5ba4bc1bb524e525fb6c3: Status 404 returned error can't find the container with id 68397e437c2fb3791ad659ab6abc466e5cb77e5b97a5ba4bc1bb524e525fb6c3 Mar 12 13:34:35 crc kubenswrapper[4778]: I0312 13:34:35.139839 4778 generic.go:334] "Generic (PLEG): container finished" podID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerID="1ca532aa466af7c68cb8aa187e7cf3ea161e9610dcf97d902b18dad6b9250f81" exitCode=0 Mar 12 13:34:35 crc kubenswrapper[4778]: I0312 13:34:35.140562 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerDied","Data":"1ca532aa466af7c68cb8aa187e7cf3ea161e9610dcf97d902b18dad6b9250f81"} Mar 12 13:34:35 crc kubenswrapper[4778]: I0312 13:34:35.140594 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerStarted","Data":"68397e437c2fb3791ad659ab6abc466e5cb77e5b97a5ba4bc1bb524e525fb6c3"} Mar 12 13:34:37 crc kubenswrapper[4778]: I0312 13:34:37.168522 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerStarted","Data":"c66167331bd74d9b577eb48b304f2f99e28d6904a5ae9cd088d4f17df80842e1"} Mar 12 13:34:38 crc kubenswrapper[4778]: I0312 13:34:38.177580 4778 generic.go:334] "Generic (PLEG): container finished" podID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerID="c66167331bd74d9b577eb48b304f2f99e28d6904a5ae9cd088d4f17df80842e1" exitCode=0 Mar 12 13:34:38 crc kubenswrapper[4778]: I0312 13:34:38.177682 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerDied","Data":"c66167331bd74d9b577eb48b304f2f99e28d6904a5ae9cd088d4f17df80842e1"} Mar 12 13:34:39 crc kubenswrapper[4778]: I0312 13:34:39.263503 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 12 13:34:40 crc kubenswrapper[4778]: I0312 13:34:40.198768 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerStarted","Data":"8dce37445b314b16965ae024d78bbfd9bf5998d5da6305572acf12733671bc3d"} Mar 12 13:34:40 crc kubenswrapper[4778]: I0312 13:34:40.223984 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jfzqk" podStartSLOduration=3.024149769 podStartE2EDuration="7.22395417s" podCreationTimestamp="2026-03-12 13:34:33 +0000 UTC" firstStartedPulling="2026-03-12 13:34:35.142746347 +0000 UTC m=+1493.591441743" lastFinishedPulling="2026-03-12 13:34:39.342550748 +0000 UTC m=+1497.791246144" observedRunningTime="2026-03-12 13:34:40.220747999 +0000 UTC m=+1498.669443405" watchObservedRunningTime="2026-03-12 13:34:40.22395417 +0000 UTC m=+1498.672649596" Mar 12 13:34:41 crc kubenswrapper[4778]: I0312 13:34:41.562554 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.066781 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qqx6r"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.068749 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.072864 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.074654 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.098868 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qqx6r"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.190778 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpk2x\" (UniqueName: \"kubernetes.io/projected/98a74774-1415-43d1-b278-bead87ab4385-kube-api-access-zpk2x\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.190861 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.192144 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-config-data\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.192205 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-scripts\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.326333 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpk2x\" (UniqueName: \"kubernetes.io/projected/98a74774-1415-43d1-b278-bead87ab4385-kube-api-access-zpk2x\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.326670 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.327207 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-config-data\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.327253 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-scripts\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.344883 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.345101 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.386356 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.397835 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-config-data\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.400053 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpk2x\" (UniqueName: \"kubernetes.io/projected/98a74774-1415-43d1-b278-bead87ab4385-kube-api-access-zpk2x\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.410413 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-scripts\") pod \"nova-cell0-cell-mapping-qqx6r\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.412999 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.487471 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.488640 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.508597 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.517039 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.557550 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.558749 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.565683 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.577822 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.617319 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.619306 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.628564 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.639732 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.669488 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-config-data\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.669589 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv5rr\" (UniqueName: \"kubernetes.io/projected/d42d33e8-c530-4272-90a4-f0ef9b061927-kube-api-access-hv5rr\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.669622 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps6cq\" (UniqueName: \"kubernetes.io/projected/65ad5500-0148-42d4-a597-53e265081516-kube-api-access-ps6cq\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.669639 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.669737 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.669762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.769218 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.771267 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.779645 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783283 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783337 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-logs\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783363 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783407 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-config-data\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783437 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsj8q\" (UniqueName: \"kubernetes.io/projected/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-kube-api-access-qsj8q\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783473 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv5rr\" (UniqueName: \"kubernetes.io/projected/d42d33e8-c530-4272-90a4-f0ef9b061927-kube-api-access-hv5rr\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783495 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783516 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps6cq\" (UniqueName: \"kubernetes.io/projected/65ad5500-0148-42d4-a597-53e265081516-kube-api-access-ps6cq\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783535 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.783551 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-config-data\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.802023 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.809875 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-config-data\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.810398 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.832034 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.849317 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.855849 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv5rr\" (UniqueName: \"kubernetes.io/projected/d42d33e8-c530-4272-90a4-f0ef9b061927-kube-api-access-hv5rr\") pod \"nova-cell1-novncproxy-0\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.884827 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps6cq\" (UniqueName: \"kubernetes.io/projected/65ad5500-0148-42d4-a597-53e265081516-kube-api-access-ps6cq\") pod \"nova-scheduler-0\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " pod="openstack/nova-scheduler-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887317 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887358 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887377 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-config-data\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887393 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-config-data\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887466 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-logs\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887506 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx6nx\" (UniqueName: \"kubernetes.io/projected/9eb26444-57b1-444a-ab45-586a64cd8857-kube-api-access-qx6nx\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887541 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsj8q\" (UniqueName: \"kubernetes.io/projected/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-kube-api-access-qsj8q\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.887568 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9eb26444-57b1-444a-ab45-586a64cd8857-logs\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.890402 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.890794 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-logs\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.897993 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.931221 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-config-data\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.940986 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsj8q\" (UniqueName: \"kubernetes.io/projected/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-kube-api-access-qsj8q\") pod \"nova-api-0\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " pod="openstack/nova-api-0" Mar 12 13:34:42 crc kubenswrapper[4778]: I0312 13:34:42.999670 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.001058 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9eb26444-57b1-444a-ab45-586a64cd8857-logs\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.001113 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.001135 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-config-data\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.001243 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx6nx\" (UniqueName: \"kubernetes.io/projected/9eb26444-57b1-444a-ab45-586a64cd8857-kube-api-access-qx6nx\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.001858 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9eb26444-57b1-444a-ab45-586a64cd8857-logs\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.008669 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.022251 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-xlfr7"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.023758 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.039372 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-xlfr7"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.051939 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-config-data\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.055652 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx6nx\" (UniqueName: \"kubernetes.io/projected/9eb26444-57b1-444a-ab45-586a64cd8857-kube-api-access-qx6nx\") pod \"nova-metadata-0\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.104958 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qqx6r"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.176307 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.197123 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.203658 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.203697 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fvsq\" (UniqueName: \"kubernetes.io/projected/f38c0efe-db9f-4afc-8693-0743c558d74f-kube-api-access-6fvsq\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.203716 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.203826 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-config\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.203851 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.203898 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-svc\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.250431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qqx6r" event={"ID":"98a74774-1415-43d1-b278-bead87ab4385","Type":"ContainerStarted","Data":"94415432161b66df8eaee31803fa6f28ac270cbac75c8be9dbd68e3fe9cda71c"} Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.309435 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.309491 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fvsq\" (UniqueName: \"kubernetes.io/projected/f38c0efe-db9f-4afc-8693-0743c558d74f-kube-api-access-6fvsq\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.309511 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.309578 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-config\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.309610 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.309663 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-svc\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.310863 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-svc\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.310870 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.311447 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-config\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.312014 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.312150 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.333235 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fvsq\" (UniqueName: \"kubernetes.io/projected/f38c0efe-db9f-4afc-8693-0743c558d74f-kube-api-access-6fvsq\") pod \"dnsmasq-dns-757b4f8459-xlfr7\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.361808 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.542101 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.643872 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.644751 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.653703 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.777562 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.824078 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7dlt6"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.825420 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.828212 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.829170 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.837046 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7dlt6"] Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.932365 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqndb\" (UniqueName: \"kubernetes.io/projected/58dfb2fb-928e-46de-90dd-481c91a7727c-kube-api-access-gqndb\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.932410 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-config-data\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.932445 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:43 crc kubenswrapper[4778]: I0312 13:34:43.932495 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-scripts\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.033990 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqndb\" (UniqueName: \"kubernetes.io/projected/58dfb2fb-928e-46de-90dd-481c91a7727c-kube-api-access-gqndb\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.034038 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-config-data\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.034071 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.034122 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-scripts\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.040554 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-config-data\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.092124 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.093599 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-scripts\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.095308 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqndb\" (UniqueName: \"kubernetes.io/projected/58dfb2fb-928e-46de-90dd-481c91a7727c-kube-api-access-gqndb\") pod \"nova-cell1-conductor-db-sync-7dlt6\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.142238 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.160017 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.161112 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-xlfr7"] Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.282796 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" event={"ID":"f38c0efe-db9f-4afc-8693-0743c558d74f","Type":"ContainerStarted","Data":"a01a33797f0031a4928ccc3b84c316e6cab0e859fc2dd6c0bc9cf5a06332acbb"} Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.290959 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d42d33e8-c530-4272-90a4-f0ef9b061927","Type":"ContainerStarted","Data":"190ba154912f1afd6c8afdd589f19abb7d2fb48d3910a0516eb35d087148f5e4"} Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.298615 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qqx6r" event={"ID":"98a74774-1415-43d1-b278-bead87ab4385","Type":"ContainerStarted","Data":"638395848d77320f6f4d74ca6334a62beda4c18b92408c089881a124597a1418"} Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.305731 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1","Type":"ContainerStarted","Data":"9d69b802526d361f0ba3ab145439034eaadd80c18ca540ff35313a518907cc83"} Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.308421 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9eb26444-57b1-444a-ab45-586a64cd8857","Type":"ContainerStarted","Data":"a723cfd0bedc0eb390903faec62e31e74919acbbe92cf204fd8b296e7d06b3bb"} Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.315538 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65ad5500-0148-42d4-a597-53e265081516","Type":"ContainerStarted","Data":"53204891ba93c9dcc714e4cf6732ebcf66cfe563b2c5b0d6b993dd7bb498dfcd"} Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.330348 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qqx6r" podStartSLOduration=2.330324731 podStartE2EDuration="2.330324731s" podCreationTimestamp="2026-03-12 13:34:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:44.315569773 +0000 UTC m=+1502.764265159" watchObservedRunningTime="2026-03-12 13:34:44.330324731 +0000 UTC m=+1502.779020127" Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.707662 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jfzqk" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" probeResult="failure" output=< Mar 12 13:34:44 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:34:44 crc kubenswrapper[4778]: > Mar 12 13:34:44 crc kubenswrapper[4778]: I0312 13:34:44.798297 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7dlt6"] Mar 12 13:34:45 crc kubenswrapper[4778]: I0312 13:34:45.346997 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" event={"ID":"58dfb2fb-928e-46de-90dd-481c91a7727c","Type":"ContainerStarted","Data":"a7c208f5185dc692f0ec8df98f6bb0b7b464e0a056d454057e864768b033e299"} Mar 12 13:34:45 crc kubenswrapper[4778]: I0312 13:34:45.347295 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" event={"ID":"58dfb2fb-928e-46de-90dd-481c91a7727c","Type":"ContainerStarted","Data":"6616a6464fcc9dedf4bc63acdc82e9b9e7114af17dd2023df257ae235015b89a"} Mar 12 13:34:45 crc kubenswrapper[4778]: I0312 13:34:45.355657 4778 generic.go:334] "Generic (PLEG): container finished" podID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerID="1e4e2a2aac1ba95c2fc03d3ae5822d197e179d60f0dbd976d4f6143a68eb2c2a" exitCode=0 Mar 12 13:34:45 crc kubenswrapper[4778]: I0312 13:34:45.355743 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" event={"ID":"f38c0efe-db9f-4afc-8693-0743c558d74f","Type":"ContainerDied","Data":"1e4e2a2aac1ba95c2fc03d3ae5822d197e179d60f0dbd976d4f6143a68eb2c2a"} Mar 12 13:34:45 crc kubenswrapper[4778]: I0312 13:34:45.401486 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" podStartSLOduration=2.401465444 podStartE2EDuration="2.401465444s" podCreationTimestamp="2026-03-12 13:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:45.378866503 +0000 UTC m=+1503.827561899" watchObservedRunningTime="2026-03-12 13:34:45.401465444 +0000 UTC m=+1503.850160840" Mar 12 13:34:46 crc kubenswrapper[4778]: I0312 13:34:46.681048 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:34:46 crc kubenswrapper[4778]: I0312 13:34:46.694571 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:47 crc kubenswrapper[4778]: I0312 13:34:47.871701 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:34:47 crc kubenswrapper[4778]: I0312 13:34:47.872985 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="66ed2760-88a0-4731-a0d1-52cb6cffa2b1" containerName="kube-state-metrics" containerID="cri-o://6addcbc9f6e1bd0c36c2127749a9343943bce9503688868083bfb8596a8eda94" gracePeriod=30 Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.403215 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" event={"ID":"f38c0efe-db9f-4afc-8693-0743c558d74f","Type":"ContainerStarted","Data":"e6738e925b347d28a1e722ea04cdc7d88018005b75c56a3dec09b214b5752ae1"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.404136 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.409711 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d42d33e8-c530-4272-90a4-f0ef9b061927","Type":"ContainerStarted","Data":"2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.410150 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d42d33e8-c530-4272-90a4-f0ef9b061927" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af" gracePeriod=30 Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.418283 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1","Type":"ContainerStarted","Data":"e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.418378 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1","Type":"ContainerStarted","Data":"8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.425152 4778 generic.go:334] "Generic (PLEG): container finished" podID="66ed2760-88a0-4731-a0d1-52cb6cffa2b1" containerID="6addcbc9f6e1bd0c36c2127749a9343943bce9503688868083bfb8596a8eda94" exitCode=2 Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.425240 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66ed2760-88a0-4731-a0d1-52cb6cffa2b1","Type":"ContainerDied","Data":"6addcbc9f6e1bd0c36c2127749a9343943bce9503688868083bfb8596a8eda94"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.425263 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66ed2760-88a0-4731-a0d1-52cb6cffa2b1","Type":"ContainerDied","Data":"4e9e8b87b4e8662cb5ee7f6527d7533b6383b322442ecf5f3470e33d6bb4be86"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.425293 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e9e8b87b4e8662cb5ee7f6527d7533b6383b322442ecf5f3470e33d6bb4be86" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.429861 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9eb26444-57b1-444a-ab45-586a64cd8857","Type":"ContainerStarted","Data":"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.429907 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9eb26444-57b1-444a-ab45-586a64cd8857","Type":"ContainerStarted","Data":"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.430079 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-log" containerID="cri-o://f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a" gracePeriod=30 Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.430562 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-metadata" containerID="cri-o://38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9" gracePeriod=30 Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.441797 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65ad5500-0148-42d4-a597-53e265081516","Type":"ContainerStarted","Data":"4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1"} Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.451427 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" podStartSLOduration=6.451407979 podStartE2EDuration="6.451407979s" podCreationTimestamp="2026-03-12 13:34:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:48.430133286 +0000 UTC m=+1506.878828682" watchObservedRunningTime="2026-03-12 13:34:48.451407979 +0000 UTC m=+1506.900103375" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.464635 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.477991 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.989076648 podStartE2EDuration="6.477968033s" podCreationTimestamp="2026-03-12 13:34:42 +0000 UTC" firstStartedPulling="2026-03-12 13:34:43.666146141 +0000 UTC m=+1502.114841537" lastFinishedPulling="2026-03-12 13:34:47.155037526 +0000 UTC m=+1505.603732922" observedRunningTime="2026-03-12 13:34:48.454076075 +0000 UTC m=+1506.902771471" watchObservedRunningTime="2026-03-12 13:34:48.477968033 +0000 UTC m=+1506.926663429" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.483907 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.121818192 podStartE2EDuration="6.48388871s" podCreationTimestamp="2026-03-12 13:34:42 +0000 UTC" firstStartedPulling="2026-03-12 13:34:43.79795361 +0000 UTC m=+1502.246648996" lastFinishedPulling="2026-03-12 13:34:47.160024118 +0000 UTC m=+1505.608719514" observedRunningTime="2026-03-12 13:34:48.474052212 +0000 UTC m=+1506.922747608" watchObservedRunningTime="2026-03-12 13:34:48.48388871 +0000 UTC m=+1506.932584106" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.500358 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.904665092 podStartE2EDuration="6.500336247s" podCreationTimestamp="2026-03-12 13:34:42 +0000 UTC" firstStartedPulling="2026-03-12 13:34:43.560546755 +0000 UTC m=+1502.009242141" lastFinishedPulling="2026-03-12 13:34:47.1562179 +0000 UTC m=+1505.604913296" observedRunningTime="2026-03-12 13:34:48.494250254 +0000 UTC m=+1506.942945650" watchObservedRunningTime="2026-03-12 13:34:48.500336247 +0000 UTC m=+1506.949031643" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.576557 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8s49\" (UniqueName: \"kubernetes.io/projected/66ed2760-88a0-4731-a0d1-52cb6cffa2b1-kube-api-access-m8s49\") pod \"66ed2760-88a0-4731-a0d1-52cb6cffa2b1\" (UID: \"66ed2760-88a0-4731-a0d1-52cb6cffa2b1\") " Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.593130 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ed2760-88a0-4731-a0d1-52cb6cffa2b1-kube-api-access-m8s49" (OuterVolumeSpecName: "kube-api-access-m8s49") pod "66ed2760-88a0-4731-a0d1-52cb6cffa2b1" (UID: "66ed2760-88a0-4731-a0d1-52cb6cffa2b1"). InnerVolumeSpecName "kube-api-access-m8s49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:48 crc kubenswrapper[4778]: I0312 13:34:48.680489 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8s49\" (UniqueName: \"kubernetes.io/projected/66ed2760-88a0-4731-a0d1-52cb6cffa2b1-kube-api-access-m8s49\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.026321 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.051989 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.033122702 podStartE2EDuration="7.051969014s" podCreationTimestamp="2026-03-12 13:34:42 +0000 UTC" firstStartedPulling="2026-03-12 13:34:44.142954386 +0000 UTC m=+1502.591649782" lastFinishedPulling="2026-03-12 13:34:47.161800698 +0000 UTC m=+1505.610496094" observedRunningTime="2026-03-12 13:34:48.534511456 +0000 UTC m=+1506.983206852" watchObservedRunningTime="2026-03-12 13:34:49.051969014 +0000 UTC m=+1507.500664410" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.190409 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-combined-ca-bundle\") pod \"9eb26444-57b1-444a-ab45-586a64cd8857\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.191096 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9eb26444-57b1-444a-ab45-586a64cd8857-logs\") pod \"9eb26444-57b1-444a-ab45-586a64cd8857\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.191206 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx6nx\" (UniqueName: \"kubernetes.io/projected/9eb26444-57b1-444a-ab45-586a64cd8857-kube-api-access-qx6nx\") pod \"9eb26444-57b1-444a-ab45-586a64cd8857\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.191358 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-config-data\") pod \"9eb26444-57b1-444a-ab45-586a64cd8857\" (UID: \"9eb26444-57b1-444a-ab45-586a64cd8857\") " Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.191482 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eb26444-57b1-444a-ab45-586a64cd8857-logs" (OuterVolumeSpecName: "logs") pod "9eb26444-57b1-444a-ab45-586a64cd8857" (UID: "9eb26444-57b1-444a-ab45-586a64cd8857"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.192080 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9eb26444-57b1-444a-ab45-586a64cd8857-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.197825 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb26444-57b1-444a-ab45-586a64cd8857-kube-api-access-qx6nx" (OuterVolumeSpecName: "kube-api-access-qx6nx") pod "9eb26444-57b1-444a-ab45-586a64cd8857" (UID: "9eb26444-57b1-444a-ab45-586a64cd8857"). InnerVolumeSpecName "kube-api-access-qx6nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.222329 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9eb26444-57b1-444a-ab45-586a64cd8857" (UID: "9eb26444-57b1-444a-ab45-586a64cd8857"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.238474 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-config-data" (OuterVolumeSpecName: "config-data") pod "9eb26444-57b1-444a-ab45-586a64cd8857" (UID: "9eb26444-57b1-444a-ab45-586a64cd8857"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.294001 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx6nx\" (UniqueName: \"kubernetes.io/projected/9eb26444-57b1-444a-ab45-586a64cd8857-kube-api-access-qx6nx\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.294042 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.294058 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb26444-57b1-444a-ab45-586a64cd8857-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458061 4778 generic.go:334] "Generic (PLEG): container finished" podID="9eb26444-57b1-444a-ab45-586a64cd8857" containerID="38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9" exitCode=0 Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458100 4778 generic.go:334] "Generic (PLEG): container finished" podID="9eb26444-57b1-444a-ab45-586a64cd8857" containerID="f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a" exitCode=143 Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458167 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458349 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458398 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9eb26444-57b1-444a-ab45-586a64cd8857","Type":"ContainerDied","Data":"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9"} Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458458 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9eb26444-57b1-444a-ab45-586a64cd8857","Type":"ContainerDied","Data":"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a"} Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458473 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9eb26444-57b1-444a-ab45-586a64cd8857","Type":"ContainerDied","Data":"a723cfd0bedc0eb390903faec62e31e74919acbbe92cf204fd8b296e7d06b3bb"} Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.458495 4778 scope.go:117] "RemoveContainer" containerID="38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.502541 4778 scope.go:117] "RemoveContainer" containerID="f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.525168 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.531948 4778 scope.go:117] "RemoveContainer" containerID="38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9" Mar 12 13:34:49 crc kubenswrapper[4778]: E0312 13:34:49.538419 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9\": container with ID starting with 38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9 not found: ID does not exist" containerID="38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.538480 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9"} err="failed to get container status \"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9\": rpc error: code = NotFound desc = could not find container \"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9\": container with ID starting with 38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9 not found: ID does not exist" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.538518 4778 scope.go:117] "RemoveContainer" containerID="f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a" Mar 12 13:34:49 crc kubenswrapper[4778]: E0312 13:34:49.544717 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a\": container with ID starting with f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a not found: ID does not exist" containerID="f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.544783 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a"} err="failed to get container status \"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a\": rpc error: code = NotFound desc = could not find container \"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a\": container with ID starting with f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a not found: ID does not exist" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.544820 4778 scope.go:117] "RemoveContainer" containerID="38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.545234 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9"} err="failed to get container status \"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9\": rpc error: code = NotFound desc = could not find container \"38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9\": container with ID starting with 38fa7ed3342fe5fa41b70791a1955f377980bfa175801ba31d13d4c89b9c16d9 not found: ID does not exist" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.545256 4778 scope.go:117] "RemoveContainer" containerID="f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.545501 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a"} err="failed to get container status \"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a\": rpc error: code = NotFound desc = could not find container \"f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a\": container with ID starting with f9211b5ee91843df422010d0e1b0d25d76aa11968301fe1adb494610b728184a not found: ID does not exist" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.550552 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.570208 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.586321 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.593692 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: E0312 13:34:49.594205 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ed2760-88a0-4731-a0d1-52cb6cffa2b1" containerName="kube-state-metrics" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.594226 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ed2760-88a0-4731-a0d1-52cb6cffa2b1" containerName="kube-state-metrics" Mar 12 13:34:49 crc kubenswrapper[4778]: E0312 13:34:49.594239 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-log" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.594247 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-log" Mar 12 13:34:49 crc kubenswrapper[4778]: E0312 13:34:49.594264 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-metadata" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.594270 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-metadata" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.594501 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-log" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.594525 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" containerName="nova-metadata-metadata" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.594550 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ed2760-88a0-4731-a0d1-52cb6cffa2b1" containerName="kube-state-metrics" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.595420 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.598683 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.598712 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.602777 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.605302 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.610150 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.610660 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.618453 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.631860 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702542 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702606 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gppmh\" (UniqueName: \"kubernetes.io/projected/51f24fcd-aff5-4785-abf7-4936180cee78-kube-api-access-gppmh\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702630 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702663 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8c2c79-f773-4580-ad43-3dcbfced2f86-logs\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702712 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702730 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702756 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702785 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-config-data\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.702813 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmpvk\" (UniqueName: \"kubernetes.io/projected/cf8c2c79-f773-4580-ad43-3dcbfced2f86-kube-api-access-pmpvk\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.806395 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.806796 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-config-data\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.806859 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmpvk\" (UniqueName: \"kubernetes.io/projected/cf8c2c79-f773-4580-ad43-3dcbfced2f86-kube-api-access-pmpvk\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.806945 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.806991 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gppmh\" (UniqueName: \"kubernetes.io/projected/51f24fcd-aff5-4785-abf7-4936180cee78-kube-api-access-gppmh\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.807020 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.807068 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8c2c79-f773-4580-ad43-3dcbfced2f86-logs\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.807139 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.807165 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.808755 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8c2c79-f773-4580-ad43-3dcbfced2f86-logs\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.814599 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.815745 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.816121 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-config-data\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.820828 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.837538 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.837968 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f24fcd-aff5-4785-abf7-4936180cee78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.842111 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gppmh\" (UniqueName: \"kubernetes.io/projected/51f24fcd-aff5-4785-abf7-4936180cee78-kube-api-access-gppmh\") pod \"kube-state-metrics-0\" (UID: \"51f24fcd-aff5-4785-abf7-4936180cee78\") " pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.842654 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmpvk\" (UniqueName: \"kubernetes.io/projected/cf8c2c79-f773-4580-ad43-3dcbfced2f86-kube-api-access-pmpvk\") pod \"nova-metadata-0\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " pod="openstack/nova-metadata-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.925139 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 13:34:49 crc kubenswrapper[4778]: I0312 13:34:49.937305 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.270419 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66ed2760-88a0-4731-a0d1-52cb6cffa2b1" path="/var/lib/kubelet/pods/66ed2760-88a0-4731-a0d1-52cb6cffa2b1/volumes" Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.271662 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb26444-57b1-444a-ab45-586a64cd8857" path="/var/lib/kubelet/pods/9eb26444-57b1-444a-ab45-586a64cd8857/volumes" Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.392018 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.392360 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-central-agent" containerID="cri-o://07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67" gracePeriod=30 Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.392483 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="sg-core" containerID="cri-o://90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c" gracePeriod=30 Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.392547 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="proxy-httpd" containerID="cri-o://8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf" gracePeriod=30 Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.392581 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-notification-agent" containerID="cri-o://5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d" gracePeriod=30 Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.435526 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.470609 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"51f24fcd-aff5-4785-abf7-4936180cee78","Type":"ContainerStarted","Data":"430dedef3bc6e9b116b7b20f0e7104500525c4e510ad2304e79eb76ac3410d4f"} Mar 12 13:34:50 crc kubenswrapper[4778]: I0312 13:34:50.512151 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.484489 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf8c2c79-f773-4580-ad43-3dcbfced2f86","Type":"ContainerStarted","Data":"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.485003 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf8c2c79-f773-4580-ad43-3dcbfced2f86","Type":"ContainerStarted","Data":"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.485023 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf8c2c79-f773-4580-ad43-3dcbfced2f86","Type":"ContainerStarted","Data":"4ec39d8d5b0fe3157074d759d6c8d58bd3fe2afde569fbe34129a0aeb9260cbc"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.487915 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"51f24fcd-aff5-4785-abf7-4936180cee78","Type":"ContainerStarted","Data":"52f48406d3459ce07c9d5861f0254b9fe6e02c6ebb107cd5294d3822b58e596f"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.488089 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.494702 4778 generic.go:334] "Generic (PLEG): container finished" podID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerID="8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf" exitCode=0 Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.494734 4778 generic.go:334] "Generic (PLEG): container finished" podID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerID="90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c" exitCode=2 Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.494743 4778 generic.go:334] "Generic (PLEG): container finished" podID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerID="07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67" exitCode=0 Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.494767 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerDied","Data":"8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.494792 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerDied","Data":"90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.494802 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerDied","Data":"07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67"} Mar 12 13:34:51 crc kubenswrapper[4778]: I0312 13:34:51.513305 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.513288372 podStartE2EDuration="2.513288372s" podCreationTimestamp="2026-03-12 13:34:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:51.504875763 +0000 UTC m=+1509.953571159" watchObservedRunningTime="2026-03-12 13:34:51.513288372 +0000 UTC m=+1509.961983768" Mar 12 13:34:52 crc kubenswrapper[4778]: I0312 13:34:52.503958 4778 generic.go:334] "Generic (PLEG): container finished" podID="98a74774-1415-43d1-b278-bead87ab4385" containerID="638395848d77320f6f4d74ca6334a62beda4c18b92408c089881a124597a1418" exitCode=0 Mar 12 13:34:52 crc kubenswrapper[4778]: I0312 13:34:52.504034 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qqx6r" event={"ID":"98a74774-1415-43d1-b278-bead87ab4385","Type":"ContainerDied","Data":"638395848d77320f6f4d74ca6334a62beda4c18b92408c089881a124597a1418"} Mar 12 13:34:52 crc kubenswrapper[4778]: I0312 13:34:52.529080 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.94125185 podStartE2EDuration="3.529054314s" podCreationTimestamp="2026-03-12 13:34:49 +0000 UTC" firstStartedPulling="2026-03-12 13:34:50.454529029 +0000 UTC m=+1508.903224425" lastFinishedPulling="2026-03-12 13:34:51.042331493 +0000 UTC m=+1509.491026889" observedRunningTime="2026-03-12 13:34:51.533773063 +0000 UTC m=+1509.982468459" watchObservedRunningTime="2026-03-12 13:34:52.529054314 +0000 UTC m=+1510.977749710" Mar 12 13:34:52 crc kubenswrapper[4778]: I0312 13:34:52.891497 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.000864 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.000921 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.176446 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.176493 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.213626 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.364417 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.434173 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-246x7"] Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.434438 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerName="dnsmasq-dns" containerID="cri-o://3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549" gracePeriod=10 Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.519589 4778 generic.go:334] "Generic (PLEG): container finished" podID="58dfb2fb-928e-46de-90dd-481c91a7727c" containerID="a7c208f5185dc692f0ec8df98f6bb0b7b464e0a056d454057e864768b033e299" exitCode=0 Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.519784 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" event={"ID":"58dfb2fb-928e-46de-90dd-481c91a7727c","Type":"ContainerDied","Data":"a7c208f5185dc692f0ec8df98f6bb0b7b464e0a056d454057e864768b033e299"} Mar 12 13:34:53 crc kubenswrapper[4778]: I0312 13:34:53.559614 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.085385 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.085407 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.103170 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.231252 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.274051 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-sb\") pod \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.274117 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-swift-storage-0\") pod \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.274200 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-svc\") pod \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.274237 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-nb\") pod \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.274272 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-config\") pod \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.274306 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62md\" (UniqueName: \"kubernetes.io/projected/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-kube-api-access-m62md\") pod \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\" (UID: \"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.326631 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-kube-api-access-m62md" (OuterVolumeSpecName: "kube-api-access-m62md") pod "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" (UID: "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3"). InnerVolumeSpecName "kube-api-access-m62md". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.369454 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" (UID: "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.384150 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpk2x\" (UniqueName: \"kubernetes.io/projected/98a74774-1415-43d1-b278-bead87ab4385-kube-api-access-zpk2x\") pod \"98a74774-1415-43d1-b278-bead87ab4385\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.384369 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-config-data\") pod \"98a74774-1415-43d1-b278-bead87ab4385\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.384833 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-scripts\") pod \"98a74774-1415-43d1-b278-bead87ab4385\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.385009 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-combined-ca-bundle\") pod \"98a74774-1415-43d1-b278-bead87ab4385\" (UID: \"98a74774-1415-43d1-b278-bead87ab4385\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.385873 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.385894 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62md\" (UniqueName: \"kubernetes.io/projected/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-kube-api-access-m62md\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.392806 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a74774-1415-43d1-b278-bead87ab4385-kube-api-access-zpk2x" (OuterVolumeSpecName: "kube-api-access-zpk2x") pod "98a74774-1415-43d1-b278-bead87ab4385" (UID: "98a74774-1415-43d1-b278-bead87ab4385"). InnerVolumeSpecName "kube-api-access-zpk2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.407387 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-scripts" (OuterVolumeSpecName: "scripts") pod "98a74774-1415-43d1-b278-bead87ab4385" (UID: "98a74774-1415-43d1-b278-bead87ab4385"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.408811 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.432919 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" (UID: "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.450175 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98a74774-1415-43d1-b278-bead87ab4385" (UID: "98a74774-1415-43d1-b278-bead87ab4385"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.466903 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" (UID: "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.468018 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" (UID: "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.469950 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-config" (OuterVolumeSpecName: "config") pod "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" (UID: "43eb6e2e-19ca-402f-a4fa-3b567ef9aef3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.479665 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-config-data" (OuterVolumeSpecName: "config-data") pod "98a74774-1415-43d1-b278-bead87ab4385" (UID: "98a74774-1415-43d1-b278-bead87ab4385"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.489364 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft9qb\" (UniqueName: \"kubernetes.io/projected/f7a2b0f7-9321-4f29-aa01-0acbc528f757-kube-api-access-ft9qb\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.489475 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-sg-core-conf-yaml\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.489662 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-run-httpd\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.489772 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-config-data\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.489980 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-log-httpd\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490078 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-combined-ca-bundle\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490144 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-scripts\") pod \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\" (UID: \"f7a2b0f7-9321-4f29-aa01-0acbc528f757\") " Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490351 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490538 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490888 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490919 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490932 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490945 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpk2x\" (UniqueName: \"kubernetes.io/projected/98a74774-1415-43d1-b278-bead87ab4385-kube-api-access-zpk2x\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490957 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490970 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490979 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7a2b0f7-9321-4f29-aa01-0acbc528f757-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490987 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.490998 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a74774-1415-43d1-b278-bead87ab4385-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.491011 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.494042 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a2b0f7-9321-4f29-aa01-0acbc528f757-kube-api-access-ft9qb" (OuterVolumeSpecName: "kube-api-access-ft9qb") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "kube-api-access-ft9qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.497555 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-scripts" (OuterVolumeSpecName: "scripts") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.532052 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.545054 4778 generic.go:334] "Generic (PLEG): container finished" podID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerID="3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549" exitCode=0 Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.545401 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" event={"ID":"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3","Type":"ContainerDied","Data":"3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549"} Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.545431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" event={"ID":"43eb6e2e-19ca-402f-a4fa-3b567ef9aef3","Type":"ContainerDied","Data":"60f6f77084cfe6904eb9dc78f60c8b66e7fa89e1a236dd4007f1375a76319d3b"} Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.545448 4778 scope.go:117] "RemoveContainer" containerID="3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.545581 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-246x7" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.574257 4778 generic.go:334] "Generic (PLEG): container finished" podID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerID="5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d" exitCode=0 Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.574336 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerDied","Data":"5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d"} Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.574369 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7a2b0f7-9321-4f29-aa01-0acbc528f757","Type":"ContainerDied","Data":"b0892c0b95fcd75a23c8ced61d0a214155429e8763fde0aacd038b1f9445ab5d"} Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.574500 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.579582 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qqx6r" event={"ID":"98a74774-1415-43d1-b278-bead87ab4385","Type":"ContainerDied","Data":"94415432161b66df8eaee31803fa6f28ac270cbac75c8be9dbd68e3fe9cda71c"} Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.579641 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94415432161b66df8eaee31803fa6f28ac270cbac75c8be9dbd68e3fe9cda71c" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.579852 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qqx6r" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.593478 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.593543 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft9qb\" (UniqueName: \"kubernetes.io/projected/f7a2b0f7-9321-4f29-aa01-0acbc528f757-kube-api-access-ft9qb\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.593561 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.638208 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.657899 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-config-data" (OuterVolumeSpecName: "config-data") pod "f7a2b0f7-9321-4f29-aa01-0acbc528f757" (UID: "f7a2b0f7-9321-4f29-aa01-0acbc528f757"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.694834 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.694886 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a2b0f7-9321-4f29-aa01-0acbc528f757-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.715769 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jfzqk" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" probeResult="failure" output=< Mar 12 13:34:54 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:34:54 crc kubenswrapper[4778]: > Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.720091 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.720335 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-log" containerID="cri-o://8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d" gracePeriod=30 Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.720750 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-api" containerID="cri-o://e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e" gracePeriod=30 Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.771492 4778 scope.go:117] "RemoveContainer" containerID="3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.830935 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-246x7"] Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.860093 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-246x7"] Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.878988 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.879309 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-log" containerID="cri-o://bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db" gracePeriod=30 Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.879381 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-metadata" containerID="cri-o://5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e" gracePeriod=30 Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.889572 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.938773 4778 scope.go:117] "RemoveContainer" containerID="3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549" Mar 12 13:34:54 crc kubenswrapper[4778]: E0312 13:34:54.940815 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549\": container with ID starting with 3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549 not found: ID does not exist" containerID="3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.940860 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549"} err="failed to get container status \"3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549\": rpc error: code = NotFound desc = could not find container \"3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549\": container with ID starting with 3acaaf246e65843557136744d8e37d230106fc2f8c2711770c3619615eeab549 not found: ID does not exist" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.940885 4778 scope.go:117] "RemoveContainer" containerID="3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31" Mar 12 13:34:54 crc kubenswrapper[4778]: E0312 13:34:54.946445 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31\": container with ID starting with 3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31 not found: ID does not exist" containerID="3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.946510 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31"} err="failed to get container status \"3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31\": rpc error: code = NotFound desc = could not find container \"3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31\": container with ID starting with 3be056ef8a27b7c5eec8e8d97597ee2f4dfeb1235b2a01b4f17cb1cb7e9cfd31 not found: ID does not exist" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.946568 4778 scope.go:117] "RemoveContainer" containerID="8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf" Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.961592 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:54 crc kubenswrapper[4778]: I0312 13:34:54.974371 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.000942 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.008211 4778 scope.go:117] "RemoveContainer" containerID="90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.046364 4778 scope.go:117] "RemoveContainer" containerID="5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.055782 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.056554 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="sg-core" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.056653 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="sg-core" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.056993 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-central-agent" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.057109 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-central-agent" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.057325 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="proxy-httpd" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.057431 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="proxy-httpd" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.057554 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a74774-1415-43d1-b278-bead87ab4385" containerName="nova-manage" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.057635 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a74774-1415-43d1-b278-bead87ab4385" containerName="nova-manage" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.057714 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58dfb2fb-928e-46de-90dd-481c91a7727c" containerName="nova-cell1-conductor-db-sync" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.057818 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="58dfb2fb-928e-46de-90dd-481c91a7727c" containerName="nova-cell1-conductor-db-sync" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.057907 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerName="init" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.058194 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerName="init" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.058847 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerName="dnsmasq-dns" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.058936 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerName="dnsmasq-dns" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.059038 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-notification-agent" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.059582 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-notification-agent" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.060011 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="sg-core" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.060131 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" containerName="dnsmasq-dns" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.060241 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="proxy-httpd" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.060367 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-notification-agent" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.060461 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" containerName="ceilometer-central-agent" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.060555 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="98a74774-1415-43d1-b278-bead87ab4385" containerName="nova-manage" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.062175 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="58dfb2fb-928e-46de-90dd-481c91a7727c" containerName="nova-cell1-conductor-db-sync" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.065092 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.072644 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.074901 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.075096 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.083141 4778 scope.go:117] "RemoveContainer" containerID="07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.084344 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.109331 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqndb\" (UniqueName: \"kubernetes.io/projected/58dfb2fb-928e-46de-90dd-481c91a7727c-kube-api-access-gqndb\") pod \"58dfb2fb-928e-46de-90dd-481c91a7727c\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.109384 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-config-data\") pod \"58dfb2fb-928e-46de-90dd-481c91a7727c\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.109467 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-combined-ca-bundle\") pod \"58dfb2fb-928e-46de-90dd-481c91a7727c\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.109544 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-scripts\") pod \"58dfb2fb-928e-46de-90dd-481c91a7727c\" (UID: \"58dfb2fb-928e-46de-90dd-481c91a7727c\") " Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110077 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110152 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82xnp\" (UniqueName: \"kubernetes.io/projected/e1488e83-3a44-41ad-aa96-de09b662c16e-kube-api-access-82xnp\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110288 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-scripts\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110319 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110361 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110386 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-run-httpd\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110432 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-log-httpd\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.110455 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-config-data\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.116455 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58dfb2fb-928e-46de-90dd-481c91a7727c-kube-api-access-gqndb" (OuterVolumeSpecName: "kube-api-access-gqndb") pod "58dfb2fb-928e-46de-90dd-481c91a7727c" (UID: "58dfb2fb-928e-46de-90dd-481c91a7727c"). InnerVolumeSpecName "kube-api-access-gqndb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.118862 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-scripts" (OuterVolumeSpecName: "scripts") pod "58dfb2fb-928e-46de-90dd-481c91a7727c" (UID: "58dfb2fb-928e-46de-90dd-481c91a7727c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.145325 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-config-data" (OuterVolumeSpecName: "config-data") pod "58dfb2fb-928e-46de-90dd-481c91a7727c" (UID: "58dfb2fb-928e-46de-90dd-481c91a7727c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.148432 4778 scope.go:117] "RemoveContainer" containerID="8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.148924 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf\": container with ID starting with 8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf not found: ID does not exist" containerID="8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.148968 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf"} err="failed to get container status \"8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf\": rpc error: code = NotFound desc = could not find container \"8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf\": container with ID starting with 8982aa7ca4022874e570ac6c59742be94301e2efa10c45e382ecf26ed4330ecf not found: ID does not exist" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.148999 4778 scope.go:117] "RemoveContainer" containerID="90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.149396 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c\": container with ID starting with 90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c not found: ID does not exist" containerID="90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.149495 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c"} err="failed to get container status \"90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c\": rpc error: code = NotFound desc = could not find container \"90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c\": container with ID starting with 90e25ee50c06dc750c5cd92c70ddda8f57bcd2a0439070e9a3f541bb4ee1e11c not found: ID does not exist" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.149585 4778 scope.go:117] "RemoveContainer" containerID="5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.149794 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58dfb2fb-928e-46de-90dd-481c91a7727c" (UID: "58dfb2fb-928e-46de-90dd-481c91a7727c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.149914 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d\": container with ID starting with 5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d not found: ID does not exist" containerID="5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.149941 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d"} err="failed to get container status \"5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d\": rpc error: code = NotFound desc = could not find container \"5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d\": container with ID starting with 5a062ded3dfdf7e0b93cfe2d1cae5ba57a787eba6993d1798acc3431826d3e6d not found: ID does not exist" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.149962 4778 scope.go:117] "RemoveContainer" containerID="07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.150363 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67\": container with ID starting with 07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67 not found: ID does not exist" containerID="07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.150431 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67"} err="failed to get container status \"07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67\": rpc error: code = NotFound desc = could not find container \"07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67\": container with ID starting with 07b0c82b03265af2d0340c98ef0951f73004b6ed91b6e9f4e0518b57f5492a67 not found: ID does not exist" Mar 12 13:34:55 crc kubenswrapper[4778]: E0312 13:34:55.171593 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4697dd3e_2fbd_4855_819e_bdd6f0d9cfe1.slice/crio-8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4697dd3e_2fbd_4855_819e_bdd6f0d9cfe1.slice/crio-conmon-8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43eb6e2e_19ca_402f_a4fa_3b567ef9aef3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98a74774_1415_43d1_b278_bead87ab4385.slice/crio-94415432161b66df8eaee31803fa6f28ac270cbac75c8be9dbd68e3fe9cda71c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43eb6e2e_19ca_402f_a4fa_3b567ef9aef3.slice/crio-60f6f77084cfe6904eb9dc78f60c8b66e7fa89e1a236dd4007f1375a76319d3b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98a74774_1415_43d1_b278_bead87ab4385.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf8c2c79_f773_4580_ad43_3dcbfced2f86.slice/crio-bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf8c2c79_f773_4580_ad43_3dcbfced2f86.slice/crio-5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e.scope\": RecentStats: unable to find data in memory cache]" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214540 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-scripts\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214608 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214660 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214681 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-run-httpd\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214721 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-log-httpd\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214740 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-config-data\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214775 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214824 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82xnp\" (UniqueName: \"kubernetes.io/projected/e1488e83-3a44-41ad-aa96-de09b662c16e-kube-api-access-82xnp\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214905 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqndb\" (UniqueName: \"kubernetes.io/projected/58dfb2fb-928e-46de-90dd-481c91a7727c-kube-api-access-gqndb\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214918 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214931 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.214941 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58dfb2fb-928e-46de-90dd-481c91a7727c-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.215751 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-run-httpd\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.215921 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-log-httpd\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.221417 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-scripts\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.221599 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-config-data\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.225741 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.227346 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.227955 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:55 crc kubenswrapper[4778]: I0312 13:34:55.229656 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82xnp\" (UniqueName: \"kubernetes.io/projected/e1488e83-3a44-41ad-aa96-de09b662c16e-kube-api-access-82xnp\") pod \"ceilometer-0\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " pod="openstack/ceilometer-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.383058 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.447240 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.523169 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-config-data\") pod \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.523338 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmpvk\" (UniqueName: \"kubernetes.io/projected/cf8c2c79-f773-4580-ad43-3dcbfced2f86-kube-api-access-pmpvk\") pod \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.523409 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-combined-ca-bundle\") pod \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.523489 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8c2c79-f773-4580-ad43-3dcbfced2f86-logs\") pod \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.523527 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-nova-metadata-tls-certs\") pod \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\" (UID: \"cf8c2c79-f773-4580-ad43-3dcbfced2f86\") " Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.524592 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf8c2c79-f773-4580-ad43-3dcbfced2f86-logs" (OuterVolumeSpecName: "logs") pod "cf8c2c79-f773-4580-ad43-3dcbfced2f86" (UID: "cf8c2c79-f773-4580-ad43-3dcbfced2f86"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.530092 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf8c2c79-f773-4580-ad43-3dcbfced2f86-kube-api-access-pmpvk" (OuterVolumeSpecName: "kube-api-access-pmpvk") pod "cf8c2c79-f773-4580-ad43-3dcbfced2f86" (UID: "cf8c2c79-f773-4580-ad43-3dcbfced2f86"). InnerVolumeSpecName "kube-api-access-pmpvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.571331 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-config-data" (OuterVolumeSpecName: "config-data") pod "cf8c2c79-f773-4580-ad43-3dcbfced2f86" (UID: "cf8c2c79-f773-4580-ad43-3dcbfced2f86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.607122 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf8c2c79-f773-4580-ad43-3dcbfced2f86" (UID: "cf8c2c79-f773-4580-ad43-3dcbfced2f86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.640566 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: E0312 13:34:55.641080 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-metadata" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.641098 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-metadata" Mar 12 13:34:56 crc kubenswrapper[4778]: E0312 13:34:55.641126 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-log" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.641135 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-log" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.641345 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-metadata" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.641367 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerName="nova-metadata-log" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.642740 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.648032 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.648063 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8c2c79-f773-4580-ad43-3dcbfced2f86-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.648073 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.648082 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmpvk\" (UniqueName: \"kubernetes.io/projected/cf8c2c79-f773-4580-ad43-3dcbfced2f86-kube-api-access-pmpvk\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.666497 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.691693 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.692071 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7dlt6" event={"ID":"58dfb2fb-928e-46de-90dd-481c91a7727c","Type":"ContainerDied","Data":"6616a6464fcc9dedf4bc63acdc82e9b9e7114af17dd2023df257ae235015b89a"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.692376 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6616a6464fcc9dedf4bc63acdc82e9b9e7114af17dd2023df257ae235015b89a" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.692389 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "cf8c2c79-f773-4580-ad43-3dcbfced2f86" (UID: "cf8c2c79-f773-4580-ad43-3dcbfced2f86"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.729798 4778 generic.go:334] "Generic (PLEG): container finished" podID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerID="5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e" exitCode=0 Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.729836 4778 generic.go:334] "Generic (PLEG): container finished" podID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" containerID="bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db" exitCode=143 Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.729903 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf8c2c79-f773-4580-ad43-3dcbfced2f86","Type":"ContainerDied","Data":"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.729930 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf8c2c79-f773-4580-ad43-3dcbfced2f86","Type":"ContainerDied","Data":"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.729940 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf8c2c79-f773-4580-ad43-3dcbfced2f86","Type":"ContainerDied","Data":"4ec39d8d5b0fe3157074d759d6c8d58bd3fe2afde569fbe34129a0aeb9260cbc"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.729954 4778 scope.go:117] "RemoveContainer" containerID="5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.730093 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.754471 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srnlz\" (UniqueName: \"kubernetes.io/projected/e28e8bc2-4b60-447e-b78e-99f53f0559e9-kube-api-access-srnlz\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.754984 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.755368 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.755795 4778 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8c2c79-f773-4580-ad43-3dcbfced2f86-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.756674 4778 generic.go:334] "Generic (PLEG): container finished" podID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerID="8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d" exitCode=143 Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.756752 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1","Type":"ContainerDied","Data":"8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.758087 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="65ad5500-0148-42d4-a597-53e265081516" containerName="nova-scheduler-scheduler" containerID="cri-o://4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" gracePeriod=30 Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.792659 4778 scope.go:117] "RemoveContainer" containerID="bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.820703 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.828569 4778 scope.go:117] "RemoveContainer" containerID="5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e" Mar 12 13:34:56 crc kubenswrapper[4778]: E0312 13:34:55.839971 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e\": container with ID starting with 5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e not found: ID does not exist" containerID="5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.840018 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e"} err="failed to get container status \"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e\": rpc error: code = NotFound desc = could not find container \"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e\": container with ID starting with 5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e not found: ID does not exist" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.840043 4778 scope.go:117] "RemoveContainer" containerID="bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db" Mar 12 13:34:56 crc kubenswrapper[4778]: E0312 13:34:55.843228 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db\": container with ID starting with bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db not found: ID does not exist" containerID="bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.843250 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db"} err="failed to get container status \"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db\": rpc error: code = NotFound desc = could not find container \"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db\": container with ID starting with bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db not found: ID does not exist" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.843266 4778 scope.go:117] "RemoveContainer" containerID="5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.843656 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e"} err="failed to get container status \"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e\": rpc error: code = NotFound desc = could not find container \"5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e\": container with ID starting with 5f8a4b137c2d402a6e035d8c3ee7d11f7df1ef398c865204e7eeb1039c06313e not found: ID does not exist" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.843690 4778 scope.go:117] "RemoveContainer" containerID="bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.843952 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db"} err="failed to get container status \"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db\": rpc error: code = NotFound desc = could not find container \"bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db\": container with ID starting with bef232eba49477a7b76b95769657b7c70c9d288b1dd88486202bf1d8cbd9a8db not found: ID does not exist" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.857217 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srnlz\" (UniqueName: \"kubernetes.io/projected/e28e8bc2-4b60-447e-b78e-99f53f0559e9-kube-api-access-srnlz\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.857255 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.857354 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.869051 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.872778 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.881052 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.885824 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srnlz\" (UniqueName: \"kubernetes.io/projected/e28e8bc2-4b60-447e-b78e-99f53f0559e9-kube-api-access-srnlz\") pod \"nova-cell1-conductor-0\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.891915 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.895340 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.899693 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.901576 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.919709 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.960733 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-config-data\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.960959 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2700355-e048-4458-b430-8d149a08d624-logs\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.961040 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.961292 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:55.961389 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbjl2\" (UniqueName: \"kubernetes.io/projected/c2700355-e048-4458-b430-8d149a08d624-kube-api-access-jbjl2\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.062927 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2700355-e048-4458-b430-8d149a08d624-logs\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.062977 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.063039 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.063072 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbjl2\" (UniqueName: \"kubernetes.io/projected/c2700355-e048-4458-b430-8d149a08d624-kube-api-access-jbjl2\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.063117 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-config-data\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.063537 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2700355-e048-4458-b430-8d149a08d624-logs\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.066847 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-config-data\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.069243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.069984 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.080812 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbjl2\" (UniqueName: \"kubernetes.io/projected/c2700355-e048-4458-b430-8d149a08d624-kube-api-access-jbjl2\") pod \"nova-metadata-0\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.088904 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.232498 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.275682 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43eb6e2e-19ca-402f-a4fa-3b567ef9aef3" path="/var/lib/kubelet/pods/43eb6e2e-19ca-402f-a4fa-3b567ef9aef3/volumes" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.276686 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf8c2c79-f773-4580-ad43-3dcbfced2f86" path="/var/lib/kubelet/pods/cf8c2c79-f773-4580-ad43-3dcbfced2f86/volumes" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.277395 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a2b0f7-9321-4f29-aa01-0acbc528f757" path="/var/lib/kubelet/pods/f7a2b0f7-9321-4f29-aa01-0acbc528f757/volumes" Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.552021 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.633494 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.740614 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.791311 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2700355-e048-4458-b430-8d149a08d624","Type":"ContainerStarted","Data":"4cdf23596db0a1e92716b41df3f9c56dd37f21ba73a9653782acae39684ee3dd"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.792902 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerStarted","Data":"9b203bf5890b3e4b6703e78b53a4c6b888b8bd4da20a4a2f1d502507cc246b88"} Mar 12 13:34:56 crc kubenswrapper[4778]: I0312 13:34:56.795408 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e28e8bc2-4b60-447e-b78e-99f53f0559e9","Type":"ContainerStarted","Data":"0b4aff5eb3ee6cc75fbeaaa57c05dff4153b4e03f714593a99c2f4d9aa7da572"} Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.807990 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2700355-e048-4458-b430-8d149a08d624","Type":"ContainerStarted","Data":"18451788e6f6468b69f6150e59f0635d08ad6db357c610ae673d149c136dfeb2"} Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.808472 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2700355-e048-4458-b430-8d149a08d624","Type":"ContainerStarted","Data":"ca24f0adae0376e480d75f053859a82c49878f39def1a1831162119084f0dc4d"} Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.812622 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerStarted","Data":"61456076e8380a1062d0214a24bdfe0fa640e7ee4451d17b11add3187cfaf9ad"} Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.816623 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e28e8bc2-4b60-447e-b78e-99f53f0559e9","Type":"ContainerStarted","Data":"17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f"} Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.817440 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.830766 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.830744751 podStartE2EDuration="2.830744751s" podCreationTimestamp="2026-03-12 13:34:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:57.825465861 +0000 UTC m=+1516.274161277" watchObservedRunningTime="2026-03-12 13:34:57.830744751 +0000 UTC m=+1516.279440147" Mar 12 13:34:57 crc kubenswrapper[4778]: I0312 13:34:57.860564 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.860544857 podStartE2EDuration="2.860544857s" podCreationTimestamp="2026-03-12 13:34:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:34:57.851308065 +0000 UTC m=+1516.300003481" watchObservedRunningTime="2026-03-12 13:34:57.860544857 +0000 UTC m=+1516.309240253" Mar 12 13:34:58 crc kubenswrapper[4778]: E0312 13:34:58.178523 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:34:58 crc kubenswrapper[4778]: E0312 13:34:58.180265 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:34:58 crc kubenswrapper[4778]: E0312 13:34:58.183624 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:34:58 crc kubenswrapper[4778]: E0312 13:34:58.183684 4778 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="65ad5500-0148-42d4-a597-53e265081516" containerName="nova-scheduler-scheduler" Mar 12 13:34:58 crc kubenswrapper[4778]: I0312 13:34:58.830117 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerStarted","Data":"825e28bca3cf084ec7f1951f758972b6df54d50fc49463a251a39ebce8dc6ce1"} Mar 12 13:34:59 crc kubenswrapper[4778]: I0312 13:34:59.840073 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerStarted","Data":"5fb44ccb1d5cc41dbcf7c6e5acea797394b81866acaa080b282103d25f4131bf"} Mar 12 13:34:59 crc kubenswrapper[4778]: I0312 13:34:59.935433 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.465554 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.565828 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-combined-ca-bundle\") pod \"65ad5500-0148-42d4-a597-53e265081516\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.565891 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps6cq\" (UniqueName: \"kubernetes.io/projected/65ad5500-0148-42d4-a597-53e265081516-kube-api-access-ps6cq\") pod \"65ad5500-0148-42d4-a597-53e265081516\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.566028 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-config-data\") pod \"65ad5500-0148-42d4-a597-53e265081516\" (UID: \"65ad5500-0148-42d4-a597-53e265081516\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.571780 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad5500-0148-42d4-a597-53e265081516-kube-api-access-ps6cq" (OuterVolumeSpecName: "kube-api-access-ps6cq") pod "65ad5500-0148-42d4-a597-53e265081516" (UID: "65ad5500-0148-42d4-a597-53e265081516"). InnerVolumeSpecName "kube-api-access-ps6cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.606481 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65ad5500-0148-42d4-a597-53e265081516" (UID: "65ad5500-0148-42d4-a597-53e265081516"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.629403 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-config-data" (OuterVolumeSpecName: "config-data") pod "65ad5500-0148-42d4-a597-53e265081516" (UID: "65ad5500-0148-42d4-a597-53e265081516"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.630471 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.667200 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-logs\") pod \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.667308 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsj8q\" (UniqueName: \"kubernetes.io/projected/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-kube-api-access-qsj8q\") pod \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.667443 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-config-data\") pod \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.667534 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-combined-ca-bundle\") pod \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\" (UID: \"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1\") " Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.667849 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-logs" (OuterVolumeSpecName: "logs") pod "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" (UID: "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.668245 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.668272 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps6cq\" (UniqueName: \"kubernetes.io/projected/65ad5500-0148-42d4-a597-53e265081516-kube-api-access-ps6cq\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.668284 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.668293 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ad5500-0148-42d4-a597-53e265081516-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.671000 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-kube-api-access-qsj8q" (OuterVolumeSpecName: "kube-api-access-qsj8q") pod "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" (UID: "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1"). InnerVolumeSpecName "kube-api-access-qsj8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.703554 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" (UID: "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.712305 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-config-data" (OuterVolumeSpecName: "config-data") pod "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" (UID: "4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.769903 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.770255 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.770274 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsj8q\" (UniqueName: \"kubernetes.io/projected/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1-kube-api-access-qsj8q\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.849889 4778 generic.go:334] "Generic (PLEG): container finished" podID="65ad5500-0148-42d4-a597-53e265081516" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" exitCode=0 Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.849953 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.849962 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65ad5500-0148-42d4-a597-53e265081516","Type":"ContainerDied","Data":"4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1"} Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.850854 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65ad5500-0148-42d4-a597-53e265081516","Type":"ContainerDied","Data":"53204891ba93c9dcc714e4cf6732ebcf66cfe563b2c5b0d6b993dd7bb498dfcd"} Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.850884 4778 scope.go:117] "RemoveContainer" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.860267 4778 generic.go:334] "Generic (PLEG): container finished" podID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerID="e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e" exitCode=0 Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.860313 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1","Type":"ContainerDied","Data":"e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e"} Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.860342 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1","Type":"ContainerDied","Data":"9d69b802526d361f0ba3ab145439034eaadd80c18ca540ff35313a518907cc83"} Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.860411 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.895352 4778 scope.go:117] "RemoveContainer" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" Mar 12 13:35:00 crc kubenswrapper[4778]: E0312 13:35:00.909310 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1\": container with ID starting with 4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1 not found: ID does not exist" containerID="4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.909387 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1"} err="failed to get container status \"4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1\": rpc error: code = NotFound desc = could not find container \"4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1\": container with ID starting with 4a58f69bc959200337168fc6797ed9ced69f8a04dbe14d36ec0e69b2498fb5e1 not found: ID does not exist" Mar 12 13:35:00 crc kubenswrapper[4778]: I0312 13:35:00.909416 4778 scope.go:117] "RemoveContainer" containerID="e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.017261 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.026585 4778 scope.go:117] "RemoveContainer" containerID="8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.042630 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.061254 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.079377 4778 scope.go:117] "RemoveContainer" containerID="e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e" Mar 12 13:35:01 crc kubenswrapper[4778]: E0312 13:35:01.079836 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e\": container with ID starting with e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e not found: ID does not exist" containerID="e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.079872 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e"} err="failed to get container status \"e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e\": rpc error: code = NotFound desc = could not find container \"e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e\": container with ID starting with e10968e0aa6d0184f80649b4d85f94854a9e9ed4e143833199a6895350db927e not found: ID does not exist" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.079893 4778 scope.go:117] "RemoveContainer" containerID="8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d" Mar 12 13:35:01 crc kubenswrapper[4778]: E0312 13:35:01.081240 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d\": container with ID starting with 8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d not found: ID does not exist" containerID="8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.081262 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d"} err="failed to get container status \"8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d\": rpc error: code = NotFound desc = could not find container \"8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d\": container with ID starting with 8ff6ebf3b7b0b27c32ef14b9af9d9ad2cb5eb0cd0fcf6c338e931544b524d41d not found: ID does not exist" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.091537 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.112301 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: E0312 13:35:01.112823 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-api" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.112848 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-api" Mar 12 13:35:01 crc kubenswrapper[4778]: E0312 13:35:01.112892 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-log" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.112902 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-log" Mar 12 13:35:01 crc kubenswrapper[4778]: E0312 13:35:01.112920 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ad5500-0148-42d4-a597-53e265081516" containerName="nova-scheduler-scheduler" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.112929 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ad5500-0148-42d4-a597-53e265081516" containerName="nova-scheduler-scheduler" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.113158 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ad5500-0148-42d4-a597-53e265081516" containerName="nova-scheduler-scheduler" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.113209 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-api" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.113229 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" containerName="nova-api-log" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.114026 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.116051 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.127824 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.133052 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.141576 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.143550 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.145472 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.153641 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284143 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045050c5-d52b-4532-baa1-e7fad66cba96-logs\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284201 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xshlg\" (UniqueName: \"kubernetes.io/projected/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-kube-api-access-xshlg\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284284 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284311 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-config-data\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284331 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-config-data\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284424 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.284462 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5p7z\" (UniqueName: \"kubernetes.io/projected/045050c5-d52b-4532-baa1-e7fad66cba96-kube-api-access-r5p7z\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.386436 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.386491 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-config-data\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.386534 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-config-data\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.386651 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.386991 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5p7z\" (UniqueName: \"kubernetes.io/projected/045050c5-d52b-4532-baa1-e7fad66cba96-kube-api-access-r5p7z\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.387506 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045050c5-d52b-4532-baa1-e7fad66cba96-logs\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.387536 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xshlg\" (UniqueName: \"kubernetes.io/projected/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-kube-api-access-xshlg\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.388120 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045050c5-d52b-4532-baa1-e7fad66cba96-logs\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.392988 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.393470 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-config-data\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.405125 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-config-data\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.409803 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.410655 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5p7z\" (UniqueName: \"kubernetes.io/projected/045050c5-d52b-4532-baa1-e7fad66cba96-kube-api-access-r5p7z\") pod \"nova-api-0\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.419026 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xshlg\" (UniqueName: \"kubernetes.io/projected/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-kube-api-access-xshlg\") pod \"nova-scheduler-0\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.568785 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.594165 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.877832 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerStarted","Data":"819e42fd8accff60f320def9e9ec88d7d0b64eac8391a4dca82bd182d50ec648"} Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.878127 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:35:01 crc kubenswrapper[4778]: I0312 13:35:01.901719 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.37174239 podStartE2EDuration="7.901704677s" podCreationTimestamp="2026-03-12 13:34:54 +0000 UTC" firstStartedPulling="2026-03-12 13:34:56.56441608 +0000 UTC m=+1515.013111476" lastFinishedPulling="2026-03-12 13:35:01.094378357 +0000 UTC m=+1519.543073763" observedRunningTime="2026-03-12 13:35:01.898144036 +0000 UTC m=+1520.346839432" watchObservedRunningTime="2026-03-12 13:35:01.901704677 +0000 UTC m=+1520.350400073" Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.077870 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:02 crc kubenswrapper[4778]: W0312 13:35:02.084928 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bc6f909_0ff5_4f18_a480_fd8e6cda5e53.slice/crio-65fd579a6354c4ba6b71d144e33926dc6f3bb53ede6b497f49fd5c5be99e7ee4 WatchSource:0}: Error finding container 65fd579a6354c4ba6b71d144e33926dc6f3bb53ede6b497f49fd5c5be99e7ee4: Status 404 returned error can't find the container with id 65fd579a6354c4ba6b71d144e33926dc6f3bb53ede6b497f49fd5c5be99e7ee4 Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.186866 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.267685 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1" path="/var/lib/kubelet/pods/4697dd3e-2fbd-4855-819e-bdd6f0d9cfe1/volumes" Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.268737 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ad5500-0148-42d4-a597-53e265081516" path="/var/lib/kubelet/pods/65ad5500-0148-42d4-a597-53e265081516/volumes" Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.887512 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"045050c5-d52b-4532-baa1-e7fad66cba96","Type":"ContainerStarted","Data":"cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40"} Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.888132 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"045050c5-d52b-4532-baa1-e7fad66cba96","Type":"ContainerStarted","Data":"e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e"} Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.888160 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"045050c5-d52b-4532-baa1-e7fad66cba96","Type":"ContainerStarted","Data":"56c1ee62c0c52d6bc024a53ffbb320a0eb1a79762a1537bb2caf3aafa91e73ce"} Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.889854 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53","Type":"ContainerStarted","Data":"dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a"} Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.889907 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53","Type":"ContainerStarted","Data":"65fd579a6354c4ba6b71d144e33926dc6f3bb53ede6b497f49fd5c5be99e7ee4"} Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.915418 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.915399052 podStartE2EDuration="2.915399052s" podCreationTimestamp="2026-03-12 13:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:02.912299594 +0000 UTC m=+1521.360995020" watchObservedRunningTime="2026-03-12 13:35:02.915399052 +0000 UTC m=+1521.364094448" Mar 12 13:35:02 crc kubenswrapper[4778]: I0312 13:35:02.940252 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.940227076 podStartE2EDuration="2.940227076s" podCreationTimestamp="2026-03-12 13:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:02.930388307 +0000 UTC m=+1521.379083703" watchObservedRunningTime="2026-03-12 13:35:02.940227076 +0000 UTC m=+1521.388922472" Mar 12 13:35:04 crc kubenswrapper[4778]: I0312 13:35:04.698516 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jfzqk" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" probeResult="failure" output=< Mar 12 13:35:04 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:35:04 crc kubenswrapper[4778]: > Mar 12 13:35:06 crc kubenswrapper[4778]: I0312 13:35:06.233777 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 13:35:06 crc kubenswrapper[4778]: I0312 13:35:06.235393 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 13:35:06 crc kubenswrapper[4778]: I0312 13:35:06.569901 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 13:35:07 crc kubenswrapper[4778]: I0312 13:35:07.251150 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:07 crc kubenswrapper[4778]: I0312 13:35:07.251206 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:11 crc kubenswrapper[4778]: I0312 13:35:11.569268 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 12 13:35:11 crc kubenswrapper[4778]: I0312 13:35:11.594488 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:35:11 crc kubenswrapper[4778]: I0312 13:35:11.594585 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:35:11 crc kubenswrapper[4778]: I0312 13:35:11.617260 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 12 13:35:12 crc kubenswrapper[4778]: I0312 13:35:12.005012 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 12 13:35:12 crc kubenswrapper[4778]: I0312 13:35:12.636508 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:12 crc kubenswrapper[4778]: I0312 13:35:12.636606 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:14 crc kubenswrapper[4778]: I0312 13:35:14.232584 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 13:35:14 crc kubenswrapper[4778]: I0312 13:35:14.232635 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 13:35:14 crc kubenswrapper[4778]: I0312 13:35:14.691967 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jfzqk" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" probeResult="failure" output=< Mar 12 13:35:14 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:35:14 crc kubenswrapper[4778]: > Mar 12 13:35:15 crc kubenswrapper[4778]: I0312 13:35:15.506006 4778 scope.go:117] "RemoveContainer" containerID="c06e4e1b6c58e04407e154a6eb32ce96d2dfbf0e7e2f81409f2e784cc2f29542" Mar 12 13:35:16 crc kubenswrapper[4778]: I0312 13:35:16.237902 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 12 13:35:16 crc kubenswrapper[4778]: I0312 13:35:16.240133 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 12 13:35:16 crc kubenswrapper[4778]: I0312 13:35:16.244325 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 12 13:35:17 crc kubenswrapper[4778]: I0312 13:35:17.021919 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.818073 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.933204 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-config-data\") pod \"d42d33e8-c530-4272-90a4-f0ef9b061927\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.933308 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv5rr\" (UniqueName: \"kubernetes.io/projected/d42d33e8-c530-4272-90a4-f0ef9b061927-kube-api-access-hv5rr\") pod \"d42d33e8-c530-4272-90a4-f0ef9b061927\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.933453 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-combined-ca-bundle\") pod \"d42d33e8-c530-4272-90a4-f0ef9b061927\" (UID: \"d42d33e8-c530-4272-90a4-f0ef9b061927\") " Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.946339 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d42d33e8-c530-4272-90a4-f0ef9b061927-kube-api-access-hv5rr" (OuterVolumeSpecName: "kube-api-access-hv5rr") pod "d42d33e8-c530-4272-90a4-f0ef9b061927" (UID: "d42d33e8-c530-4272-90a4-f0ef9b061927"). InnerVolumeSpecName "kube-api-access-hv5rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.960588 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-config-data" (OuterVolumeSpecName: "config-data") pod "d42d33e8-c530-4272-90a4-f0ef9b061927" (UID: "d42d33e8-c530-4272-90a4-f0ef9b061927"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:18 crc kubenswrapper[4778]: I0312 13:35:18.963110 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d42d33e8-c530-4272-90a4-f0ef9b061927" (UID: "d42d33e8-c530-4272-90a4-f0ef9b061927"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.035330 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.035366 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42d33e8-c530-4272-90a4-f0ef9b061927-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.035377 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv5rr\" (UniqueName: \"kubernetes.io/projected/d42d33e8-c530-4272-90a4-f0ef9b061927-kube-api-access-hv5rr\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.037589 4778 generic.go:334] "Generic (PLEG): container finished" podID="d42d33e8-c530-4272-90a4-f0ef9b061927" containerID="2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af" exitCode=137 Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.037621 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.037652 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d42d33e8-c530-4272-90a4-f0ef9b061927","Type":"ContainerDied","Data":"2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af"} Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.037686 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d42d33e8-c530-4272-90a4-f0ef9b061927","Type":"ContainerDied","Data":"190ba154912f1afd6c8afdd589f19abb7d2fb48d3910a0516eb35d087148f5e4"} Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.037703 4778 scope.go:117] "RemoveContainer" containerID="2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.059594 4778 scope.go:117] "RemoveContainer" containerID="2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af" Mar 12 13:35:19 crc kubenswrapper[4778]: E0312 13:35:19.059920 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af\": container with ID starting with 2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af not found: ID does not exist" containerID="2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.060026 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af"} err="failed to get container status \"2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af\": rpc error: code = NotFound desc = could not find container \"2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af\": container with ID starting with 2daa8ef0c43b0a0e16322a9531b1ccfd1b86a58c1ab4dbd58ffa5e731b6266af not found: ID does not exist" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.075934 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.085972 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.107491 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:35:19 crc kubenswrapper[4778]: E0312 13:35:19.107939 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d42d33e8-c530-4272-90a4-f0ef9b061927" containerName="nova-cell1-novncproxy-novncproxy" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.107956 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d42d33e8-c530-4272-90a4-f0ef9b061927" containerName="nova-cell1-novncproxy-novncproxy" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.108141 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d42d33e8-c530-4272-90a4-f0ef9b061927" containerName="nova-cell1-novncproxy-novncproxy" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.108846 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.113823 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.114067 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.114323 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.118802 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.238097 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.238177 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmqmm\" (UniqueName: \"kubernetes.io/projected/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-kube-api-access-fmqmm\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.238225 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.238279 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.238303 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.340172 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmqmm\" (UniqueName: \"kubernetes.io/projected/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-kube-api-access-fmqmm\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.340268 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.340372 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.340403 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.340500 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.344149 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.344626 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.344889 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.345058 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.356163 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmqmm\" (UniqueName: \"kubernetes.io/projected/2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7-kube-api-access-fmqmm\") pod \"nova-cell1-novncproxy-0\" (UID: \"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.432284 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.594626 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.595370 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 13:35:19 crc kubenswrapper[4778]: I0312 13:35:19.850368 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 13:35:19 crc kubenswrapper[4778]: W0312 13:35:19.856816 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b43a8b1_b8bc_4ab5_af66_674fa7ff47d7.slice/crio-fb2cd569d81e6acd2dd3f3c077e284749a8ed5785ff148caac8ecbc2aef7b69c WatchSource:0}: Error finding container fb2cd569d81e6acd2dd3f3c077e284749a8ed5785ff148caac8ecbc2aef7b69c: Status 404 returned error can't find the container with id fb2cd569d81e6acd2dd3f3c077e284749a8ed5785ff148caac8ecbc2aef7b69c Mar 12 13:35:20 crc kubenswrapper[4778]: I0312 13:35:20.050166 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7","Type":"ContainerStarted","Data":"fb2cd569d81e6acd2dd3f3c077e284749a8ed5785ff148caac8ecbc2aef7b69c"} Mar 12 13:35:20 crc kubenswrapper[4778]: I0312 13:35:20.265590 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d42d33e8-c530-4272-90a4-f0ef9b061927" path="/var/lib/kubelet/pods/d42d33e8-c530-4272-90a4-f0ef9b061927/volumes" Mar 12 13:35:21 crc kubenswrapper[4778]: I0312 13:35:21.062158 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7","Type":"ContainerStarted","Data":"190d97198e753f9e9071d5fc2d25934da1a1861cf6d3a0b1fe6071cce7b119aa"} Mar 12 13:35:21 crc kubenswrapper[4778]: I0312 13:35:21.083756 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.083735891 podStartE2EDuration="2.083735891s" podCreationTimestamp="2026-03-12 13:35:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:21.076171136 +0000 UTC m=+1539.524866542" watchObservedRunningTime="2026-03-12 13:35:21.083735891 +0000 UTC m=+1539.532431307" Mar 12 13:35:21 crc kubenswrapper[4778]: I0312 13:35:21.599257 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 13:35:21 crc kubenswrapper[4778]: I0312 13:35:21.599738 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 13:35:21 crc kubenswrapper[4778]: I0312 13:35:21.602491 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.074960 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.332557 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vbzn5"] Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.356506 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vbzn5"] Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.356630 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.498892 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.499049 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.499122 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.499279 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.499330 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5ccj\" (UniqueName: \"kubernetes.io/projected/d621990b-b3fb-457c-a7b8-0726fa89a5e6-kube-api-access-b5ccj\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.499499 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-config\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.601252 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.601382 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.601426 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.601451 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5ccj\" (UniqueName: \"kubernetes.io/projected/d621990b-b3fb-457c-a7b8-0726fa89a5e6-kube-api-access-b5ccj\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.601497 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-config\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.601543 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.602784 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.602813 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.603460 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.604120 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.605294 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-config\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.626491 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5ccj\" (UniqueName: \"kubernetes.io/projected/d621990b-b3fb-457c-a7b8-0726fa89a5e6-kube-api-access-b5ccj\") pod \"dnsmasq-dns-89c5cd4d5-vbzn5\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:22 crc kubenswrapper[4778]: I0312 13:35:22.692538 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:23 crc kubenswrapper[4778]: I0312 13:35:23.188216 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vbzn5"] Mar 12 13:35:23 crc kubenswrapper[4778]: I0312 13:35:23.695602 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:35:23 crc kubenswrapper[4778]: I0312 13:35:23.773582 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:35:23 crc kubenswrapper[4778]: I0312 13:35:23.946710 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jfzqk"] Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.088383 4778 generic.go:334] "Generic (PLEG): container finished" podID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerID="f768634e6581a58404932d5b274b7e499ff8a446926b77d44c652d5c4c0bad66" exitCode=0 Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.088428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" event={"ID":"d621990b-b3fb-457c-a7b8-0726fa89a5e6","Type":"ContainerDied","Data":"f768634e6581a58404932d5b274b7e499ff8a446926b77d44c652d5c4c0bad66"} Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.088491 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" event={"ID":"d621990b-b3fb-457c-a7b8-0726fa89a5e6","Type":"ContainerStarted","Data":"7c046518ad4ee249311d20eb84f556ea55869944e1e9d121bc2b448648522cec"} Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.336652 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.336892 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-central-agent" containerID="cri-o://61456076e8380a1062d0214a24bdfe0fa640e7ee4451d17b11add3187cfaf9ad" gracePeriod=30 Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.337760 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="proxy-httpd" containerID="cri-o://819e42fd8accff60f320def9e9ec88d7d0b64eac8391a4dca82bd182d50ec648" gracePeriod=30 Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.337922 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-notification-agent" containerID="cri-o://825e28bca3cf084ec7f1951f758972b6df54d50fc49463a251a39ebce8dc6ce1" gracePeriod=30 Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.337927 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="sg-core" containerID="cri-o://5fb44ccb1d5cc41dbcf7c6e5acea797394b81866acaa080b282103d25f4131bf" gracePeriod=30 Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.358291 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.198:3000/\": EOF" Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.433161 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:24 crc kubenswrapper[4778]: I0312 13:35:24.659020 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101556 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerID="819e42fd8accff60f320def9e9ec88d7d0b64eac8391a4dca82bd182d50ec648" exitCode=0 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101595 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerID="5fb44ccb1d5cc41dbcf7c6e5acea797394b81866acaa080b282103d25f4131bf" exitCode=2 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101606 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerID="825e28bca3cf084ec7f1951f758972b6df54d50fc49463a251a39ebce8dc6ce1" exitCode=0 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101616 4778 generic.go:334] "Generic (PLEG): container finished" podID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerID="61456076e8380a1062d0214a24bdfe0fa640e7ee4451d17b11add3187cfaf9ad" exitCode=0 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101612 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerDied","Data":"819e42fd8accff60f320def9e9ec88d7d0b64eac8391a4dca82bd182d50ec648"} Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101729 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerDied","Data":"5fb44ccb1d5cc41dbcf7c6e5acea797394b81866acaa080b282103d25f4131bf"} Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101742 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerDied","Data":"825e28bca3cf084ec7f1951f758972b6df54d50fc49463a251a39ebce8dc6ce1"} Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.101753 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerDied","Data":"61456076e8380a1062d0214a24bdfe0fa640e7ee4451d17b11add3187cfaf9ad"} Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.105295 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" event={"ID":"d621990b-b3fb-457c-a7b8-0726fa89a5e6","Type":"ContainerStarted","Data":"9226d052c31f98b5c3da17ce19bbc81e718b949c212eab5fa79f7c540fdf830a"} Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.105497 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jfzqk" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" containerID="cri-o://8dce37445b314b16965ae024d78bbfd9bf5998d5da6305572acf12733671bc3d" gracePeriod=2 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.105738 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.105846 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-log" containerID="cri-o://e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e" gracePeriod=30 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.105918 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-api" containerID="cri-o://cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40" gracePeriod=30 Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.177734 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" podStartSLOduration=3.17771608 podStartE2EDuration="3.17771608s" podCreationTimestamp="2026-03-12 13:35:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:25.17667391 +0000 UTC m=+1543.625369306" watchObservedRunningTime="2026-03-12 13:35:25.17771608 +0000 UTC m=+1543.626411476" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.545994 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.563236 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-combined-ca-bundle\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.563323 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-config-data\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.563370 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82xnp\" (UniqueName: \"kubernetes.io/projected/e1488e83-3a44-41ad-aa96-de09b662c16e-kube-api-access-82xnp\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.563435 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-log-httpd\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.563467 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-sg-core-conf-yaml\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.563483 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-scripts\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.564011 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-run-httpd\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.564195 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-ceilometer-tls-certs\") pod \"e1488e83-3a44-41ad-aa96-de09b662c16e\" (UID: \"e1488e83-3a44-41ad-aa96-de09b662c16e\") " Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.564626 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.565156 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.565368 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.565443 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1488e83-3a44-41ad-aa96-de09b662c16e-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.570400 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1488e83-3a44-41ad-aa96-de09b662c16e-kube-api-access-82xnp" (OuterVolumeSpecName: "kube-api-access-82xnp") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "kube-api-access-82xnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.571694 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-scripts" (OuterVolumeSpecName: "scripts") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.624367 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.634695 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.667317 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82xnp\" (UniqueName: \"kubernetes.io/projected/e1488e83-3a44-41ad-aa96-de09b662c16e-kube-api-access-82xnp\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.667578 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.667589 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.667599 4778 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.667767 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.696099 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-config-data" (OuterVolumeSpecName: "config-data") pod "e1488e83-3a44-41ad-aa96-de09b662c16e" (UID: "e1488e83-3a44-41ad-aa96-de09b662c16e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.768850 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:25 crc kubenswrapper[4778]: I0312 13:35:25.768889 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1488e83-3a44-41ad-aa96-de09b662c16e-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.127938 4778 generic.go:334] "Generic (PLEG): container finished" podID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerID="8dce37445b314b16965ae024d78bbfd9bf5998d5da6305572acf12733671bc3d" exitCode=0 Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.128124 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerDied","Data":"8dce37445b314b16965ae024d78bbfd9bf5998d5da6305572acf12733671bc3d"} Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.128332 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzqk" event={"ID":"1b7d48c4-04cd-481a-976d-19e57a28a1d9","Type":"ContainerDied","Data":"68397e437c2fb3791ad659ab6abc466e5cb77e5b97a5ba4bc1bb524e525fb6c3"} Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.128350 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68397e437c2fb3791ad659ab6abc466e5cb77e5b97a5ba4bc1bb524e525fb6c3" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.129855 4778 generic.go:334] "Generic (PLEG): container finished" podID="045050c5-d52b-4532-baa1-e7fad66cba96" containerID="e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e" exitCode=143 Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.129900 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"045050c5-d52b-4532-baa1-e7fad66cba96","Type":"ContainerDied","Data":"e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e"} Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.132759 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.132876 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1488e83-3a44-41ad-aa96-de09b662c16e","Type":"ContainerDied","Data":"9b203bf5890b3e4b6703e78b53a4c6b888b8bd4da20a4a2f1d502507cc246b88"} Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.132905 4778 scope.go:117] "RemoveContainer" containerID="819e42fd8accff60f320def9e9ec88d7d0b64eac8391a4dca82bd182d50ec648" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.181265 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.184518 4778 scope.go:117] "RemoveContainer" containerID="5fb44ccb1d5cc41dbcf7c6e5acea797394b81866acaa080b282103d25f4131bf" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.209225 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.231596 4778 scope.go:117] "RemoveContainer" containerID="825e28bca3cf084ec7f1951f758972b6df54d50fc49463a251a39ebce8dc6ce1" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.232010 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.248068 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.248554 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="proxy-httpd" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.248576 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="proxy-httpd" Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.248591 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.248599 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.248611 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="extract-content" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249180 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="extract-content" Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.249222 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="sg-core" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249233 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="sg-core" Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.249245 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-notification-agent" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249253 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-notification-agent" Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.249288 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-central-agent" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249296 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-central-agent" Mar 12 13:35:26 crc kubenswrapper[4778]: E0312 13:35:26.249317 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="extract-utilities" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249325 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="extract-utilities" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249569 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="proxy-httpd" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249600 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-notification-agent" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249614 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="sg-core" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249629 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="ceilometer-central-agent" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.249644 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" containerName="registry-server" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.251401 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.264789 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.264978 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.266802 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.271480 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" path="/var/lib/kubelet/pods/e1488e83-3a44-41ad-aa96-de09b662c16e/volumes" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.287570 4778 scope.go:117] "RemoveContainer" containerID="61456076e8380a1062d0214a24bdfe0fa640e7ee4451d17b11add3187cfaf9ad" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.294548 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-utilities\") pod \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.294747 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-catalog-content\") pod \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.294878 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frrb7\" (UniqueName: \"kubernetes.io/projected/1b7d48c4-04cd-481a-976d-19e57a28a1d9-kube-api-access-frrb7\") pod \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\" (UID: \"1b7d48c4-04cd-481a-976d-19e57a28a1d9\") " Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295229 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-scripts\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295260 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295286 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-log-httpd\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295318 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295364 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-utilities" (OuterVolumeSpecName: "utilities") pod "1b7d48c4-04cd-481a-976d-19e57a28a1d9" (UID: "1b7d48c4-04cd-481a-976d-19e57a28a1d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295549 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-config-data\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295689 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295774 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjcm\" (UniqueName: \"kubernetes.io/projected/bbcdf243-9822-4089-9cae-4a46417b6dc0-kube-api-access-2fjcm\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.295882 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-run-httpd\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.296067 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.298672 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.304407 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b7d48c4-04cd-481a-976d-19e57a28a1d9-kube-api-access-frrb7" (OuterVolumeSpecName: "kube-api-access-frrb7") pod "1b7d48c4-04cd-481a-976d-19e57a28a1d9" (UID: "1b7d48c4-04cd-481a-976d-19e57a28a1d9"). InnerVolumeSpecName "kube-api-access-frrb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.397732 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.397783 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-log-httpd\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.397810 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.397894 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-config-data\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.397932 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.397959 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjcm\" (UniqueName: \"kubernetes.io/projected/bbcdf243-9822-4089-9cae-4a46417b6dc0-kube-api-access-2fjcm\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.398000 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-run-httpd\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.398035 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-scripts\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.398094 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frrb7\" (UniqueName: \"kubernetes.io/projected/1b7d48c4-04cd-481a-976d-19e57a28a1d9-kube-api-access-frrb7\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.398515 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-log-httpd\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.398815 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-run-httpd\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.402612 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-scripts\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.403074 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.403639 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.403992 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.409384 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-config-data\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.416992 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjcm\" (UniqueName: \"kubernetes.io/projected/bbcdf243-9822-4089-9cae-4a46417b6dc0-kube-api-access-2fjcm\") pod \"ceilometer-0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.425531 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b7d48c4-04cd-481a-976d-19e57a28a1d9" (UID: "1b7d48c4-04cd-481a-976d-19e57a28a1d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.500501 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b7d48c4-04cd-481a-976d-19e57a28a1d9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.588823 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:26 crc kubenswrapper[4778]: I0312 13:35:26.589971 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:27 crc kubenswrapper[4778]: I0312 13:35:27.012721 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:27 crc kubenswrapper[4778]: I0312 13:35:27.145482 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerStarted","Data":"e349198afaff0969683d3154a99e49c5908b20bc0714e59a3832484e545b97dc"} Mar 12 13:35:27 crc kubenswrapper[4778]: I0312 13:35:27.145527 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzqk" Mar 12 13:35:27 crc kubenswrapper[4778]: I0312 13:35:27.206864 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jfzqk"] Mar 12 13:35:27 crc kubenswrapper[4778]: I0312 13:35:27.218782 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jfzqk"] Mar 12 13:35:28 crc kubenswrapper[4778]: I0312 13:35:28.158693 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerStarted","Data":"f6f2930d9f70388763ddc5deccd561746f3634cc538e9cb6c56ef8628fd4e069"} Mar 12 13:35:28 crc kubenswrapper[4778]: I0312 13:35:28.270827 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b7d48c4-04cd-481a-976d-19e57a28a1d9" path="/var/lib/kubelet/pods/1b7d48c4-04cd-481a-976d-19e57a28a1d9/volumes" Mar 12 13:35:28 crc kubenswrapper[4778]: I0312 13:35:28.557750 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:35:28 crc kubenswrapper[4778]: I0312 13:35:28.558073 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.102376 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.150583 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045050c5-d52b-4532-baa1-e7fad66cba96-logs\") pod \"045050c5-d52b-4532-baa1-e7fad66cba96\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.151067 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-config-data\") pod \"045050c5-d52b-4532-baa1-e7fad66cba96\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.151472 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5p7z\" (UniqueName: \"kubernetes.io/projected/045050c5-d52b-4532-baa1-e7fad66cba96-kube-api-access-r5p7z\") pod \"045050c5-d52b-4532-baa1-e7fad66cba96\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.152001 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/045050c5-d52b-4532-baa1-e7fad66cba96-logs" (OuterVolumeSpecName: "logs") pod "045050c5-d52b-4532-baa1-e7fad66cba96" (UID: "045050c5-d52b-4532-baa1-e7fad66cba96"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.156427 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/045050c5-d52b-4532-baa1-e7fad66cba96-kube-api-access-r5p7z" (OuterVolumeSpecName: "kube-api-access-r5p7z") pod "045050c5-d52b-4532-baa1-e7fad66cba96" (UID: "045050c5-d52b-4532-baa1-e7fad66cba96"). InnerVolumeSpecName "kube-api-access-r5p7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.175335 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-combined-ca-bundle\") pod \"045050c5-d52b-4532-baa1-e7fad66cba96\" (UID: \"045050c5-d52b-4532-baa1-e7fad66cba96\") " Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.176384 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/045050c5-d52b-4532-baa1-e7fad66cba96-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.176407 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5p7z\" (UniqueName: \"kubernetes.io/projected/045050c5-d52b-4532-baa1-e7fad66cba96-kube-api-access-r5p7z\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.212735 4778 generic.go:334] "Generic (PLEG): container finished" podID="045050c5-d52b-4532-baa1-e7fad66cba96" containerID="cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40" exitCode=0 Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.212839 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"045050c5-d52b-4532-baa1-e7fad66cba96","Type":"ContainerDied","Data":"cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40"} Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.212868 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"045050c5-d52b-4532-baa1-e7fad66cba96","Type":"ContainerDied","Data":"56c1ee62c0c52d6bc024a53ffbb320a0eb1a79762a1537bb2caf3aafa91e73ce"} Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.212888 4778 scope.go:117] "RemoveContainer" containerID="cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.213023 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.250112 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerStarted","Data":"7114be6621c79d0d604f29d2e6499dffdde39edd34fd34dad40202ec3b0b6eef"} Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.298893 4778 scope.go:117] "RemoveContainer" containerID="e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.372526 4778 scope.go:117] "RemoveContainer" containerID="cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40" Mar 12 13:35:29 crc kubenswrapper[4778]: E0312 13:35:29.381724 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40\": container with ID starting with cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40 not found: ID does not exist" containerID="cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.381775 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40"} err="failed to get container status \"cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40\": rpc error: code = NotFound desc = could not find container \"cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40\": container with ID starting with cdf213319669e5763e5a6be5e5f3c8d41efefcfd15e81ad3ad34fb03c9028e40 not found: ID does not exist" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.381804 4778 scope.go:117] "RemoveContainer" containerID="e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e" Mar 12 13:35:29 crc kubenswrapper[4778]: E0312 13:35:29.386996 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e\": container with ID starting with e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e not found: ID does not exist" containerID="e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.387044 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e"} err="failed to get container status \"e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e\": rpc error: code = NotFound desc = could not find container \"e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e\": container with ID starting with e423b9df02af65c8d19318720fc217bb660a4ff96461ba94e43c2bb5658deb5e not found: ID does not exist" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.433441 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-config-data" (OuterVolumeSpecName: "config-data") pod "045050c5-d52b-4532-baa1-e7fad66cba96" (UID: "045050c5-d52b-4532-baa1-e7fad66cba96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.434170 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.441308 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "045050c5-d52b-4532-baa1-e7fad66cba96" (UID: "045050c5-d52b-4532-baa1-e7fad66cba96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.467943 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.485517 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.485561 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/045050c5-d52b-4532-baa1-e7fad66cba96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.546220 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.557927 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.576166 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:29 crc kubenswrapper[4778]: E0312 13:35:29.576692 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-api" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.576715 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-api" Mar 12 13:35:29 crc kubenswrapper[4778]: E0312 13:35:29.576763 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-log" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.576770 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-log" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.577015 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-api" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.577042 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" containerName="nova-api-log" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.578360 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.582583 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.582818 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.582969 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.585837 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.688588 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.688704 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-public-tls-certs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.688751 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.688796 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a186b68c-e472-4507-abc7-0b90ca321ded-logs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.688947 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-config-data\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.689300 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27cs2\" (UniqueName: \"kubernetes.io/projected/a186b68c-e472-4507-abc7-0b90ca321ded-kube-api-access-27cs2\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.791242 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a186b68c-e472-4507-abc7-0b90ca321ded-logs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.791361 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-config-data\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.791414 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27cs2\" (UniqueName: \"kubernetes.io/projected/a186b68c-e472-4507-abc7-0b90ca321ded-kube-api-access-27cs2\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.791450 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.791503 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-public-tls-certs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.792390 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.795245 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a186b68c-e472-4507-abc7-0b90ca321ded-logs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.797314 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.798533 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.798749 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-public-tls-certs\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.799879 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-config-data\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.809222 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27cs2\" (UniqueName: \"kubernetes.io/projected/a186b68c-e472-4507-abc7-0b90ca321ded-kube-api-access-27cs2\") pod \"nova-api-0\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " pod="openstack/nova-api-0" Mar 12 13:35:29 crc kubenswrapper[4778]: I0312 13:35:29.905043 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.289451 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="045050c5-d52b-4532-baa1-e7fad66cba96" path="/var/lib/kubelet/pods/045050c5-d52b-4532-baa1-e7fad66cba96/volumes" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.293723 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerStarted","Data":"8683a8e7549e2bde381c989208b511414ec56e8f866bf125984b6c4530f4d727"} Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.309047 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.424402 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:30 crc kubenswrapper[4778]: W0312 13:35:30.426400 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda186b68c_e472_4507_abc7_0b90ca321ded.slice/crio-f5dd18bc2fa3f41fb9117d4f1e4c37d3b6b9987574f509d03e4a076b1981eb8c WatchSource:0}: Error finding container f5dd18bc2fa3f41fb9117d4f1e4c37d3b6b9987574f509d03e4a076b1981eb8c: Status 404 returned error can't find the container with id f5dd18bc2fa3f41fb9117d4f1e4c37d3b6b9987574f509d03e4a076b1981eb8c Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.541329 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9xw6b"] Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.543038 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.547024 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.547347 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.549096 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9xw6b"] Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.606518 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsjt8\" (UniqueName: \"kubernetes.io/projected/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-kube-api-access-zsjt8\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.606647 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-scripts\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.606900 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-config-data\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.606967 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.708835 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-config-data\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.709099 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.709249 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsjt8\" (UniqueName: \"kubernetes.io/projected/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-kube-api-access-zsjt8\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.709489 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-scripts\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.712543 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.714367 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-scripts\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.714677 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-config-data\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.725701 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsjt8\" (UniqueName: \"kubernetes.io/projected/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-kube-api-access-zsjt8\") pod \"nova-cell1-cell-mapping-9xw6b\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:30 crc kubenswrapper[4778]: I0312 13:35:30.868143 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:31 crc kubenswrapper[4778]: I0312 13:35:31.308912 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a186b68c-e472-4507-abc7-0b90ca321ded","Type":"ContainerStarted","Data":"eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11"} Mar 12 13:35:31 crc kubenswrapper[4778]: I0312 13:35:31.309256 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a186b68c-e472-4507-abc7-0b90ca321ded","Type":"ContainerStarted","Data":"7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397"} Mar 12 13:35:31 crc kubenswrapper[4778]: I0312 13:35:31.309269 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a186b68c-e472-4507-abc7-0b90ca321ded","Type":"ContainerStarted","Data":"f5dd18bc2fa3f41fb9117d4f1e4c37d3b6b9987574f509d03e4a076b1981eb8c"} Mar 12 13:35:31 crc kubenswrapper[4778]: I0312 13:35:31.329063 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9xw6b"] Mar 12 13:35:31 crc kubenswrapper[4778]: I0312 13:35:31.342257 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.3422359520000002 podStartE2EDuration="2.342235952s" podCreationTimestamp="2026-03-12 13:35:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:31.334097801 +0000 UTC m=+1549.782793197" watchObservedRunningTime="2026-03-12 13:35:31.342235952 +0000 UTC m=+1549.790931348" Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.320259 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerStarted","Data":"e8d17472cef396ced990a10ecac98a4762149d480cd7b1355d84ce3ecdbcf8ad"} Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.320852 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.320529 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="proxy-httpd" containerID="cri-o://e8d17472cef396ced990a10ecac98a4762149d480cd7b1355d84ce3ecdbcf8ad" gracePeriod=30 Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.320322 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-central-agent" containerID="cri-o://f6f2930d9f70388763ddc5deccd561746f3634cc538e9cb6c56ef8628fd4e069" gracePeriod=30 Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.320555 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-notification-agent" containerID="cri-o://7114be6621c79d0d604f29d2e6499dffdde39edd34fd34dad40202ec3b0b6eef" gracePeriod=30 Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.320543 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="sg-core" containerID="cri-o://8683a8e7549e2bde381c989208b511414ec56e8f866bf125984b6c4530f4d727" gracePeriod=30 Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.325413 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9xw6b" event={"ID":"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3","Type":"ContainerStarted","Data":"a3547232ddc46df5ded5cc24fff2ec3e7c8bb4fb4c52277d66e27c319ec41995"} Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.325461 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9xw6b" event={"ID":"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3","Type":"ContainerStarted","Data":"fa86e251def50fe26c7890455a492370653fec6579cf29dd2f2d83fb340958c7"} Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.348971 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.639212143 podStartE2EDuration="6.348951328s" podCreationTimestamp="2026-03-12 13:35:26 +0000 UTC" firstStartedPulling="2026-03-12 13:35:27.018234598 +0000 UTC m=+1545.466929994" lastFinishedPulling="2026-03-12 13:35:31.727973783 +0000 UTC m=+1550.176669179" observedRunningTime="2026-03-12 13:35:32.343333068 +0000 UTC m=+1550.792028464" watchObservedRunningTime="2026-03-12 13:35:32.348951328 +0000 UTC m=+1550.797646724" Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.369402 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9xw6b" podStartSLOduration=2.369382017 podStartE2EDuration="2.369382017s" podCreationTimestamp="2026-03-12 13:35:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:32.359378594 +0000 UTC m=+1550.808074000" watchObservedRunningTime="2026-03-12 13:35:32.369382017 +0000 UTC m=+1550.818077413" Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.694406 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.782327 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-xlfr7"] Mar 12 13:35:32 crc kubenswrapper[4778]: I0312 13:35:32.782842 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="dnsmasq-dns" containerID="cri-o://e6738e925b347d28a1e722ea04cdc7d88018005b75c56a3dec09b214b5752ae1" gracePeriod=10 Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.339744 4778 generic.go:334] "Generic (PLEG): container finished" podID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerID="e6738e925b347d28a1e722ea04cdc7d88018005b75c56a3dec09b214b5752ae1" exitCode=0 Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.339831 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" event={"ID":"f38c0efe-db9f-4afc-8693-0743c558d74f","Type":"ContainerDied","Data":"e6738e925b347d28a1e722ea04cdc7d88018005b75c56a3dec09b214b5752ae1"} Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.340326 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" event={"ID":"f38c0efe-db9f-4afc-8693-0743c558d74f","Type":"ContainerDied","Data":"a01a33797f0031a4928ccc3b84c316e6cab0e859fc2dd6c0bc9cf5a06332acbb"} Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.340344 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01a33797f0031a4928ccc3b84c316e6cab0e859fc2dd6c0bc9cf5a06332acbb" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.353587 4778 generic.go:334] "Generic (PLEG): container finished" podID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerID="e8d17472cef396ced990a10ecac98a4762149d480cd7b1355d84ce3ecdbcf8ad" exitCode=0 Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.353651 4778 generic.go:334] "Generic (PLEG): container finished" podID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerID="8683a8e7549e2bde381c989208b511414ec56e8f866bf125984b6c4530f4d727" exitCode=2 Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.353666 4778 generic.go:334] "Generic (PLEG): container finished" podID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerID="7114be6621c79d0d604f29d2e6499dffdde39edd34fd34dad40202ec3b0b6eef" exitCode=0 Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.353672 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerDied","Data":"e8d17472cef396ced990a10ecac98a4762149d480cd7b1355d84ce3ecdbcf8ad"} Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.353725 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerDied","Data":"8683a8e7549e2bde381c989208b511414ec56e8f866bf125984b6c4530f4d727"} Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.353740 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerDied","Data":"7114be6621c79d0d604f29d2e6499dffdde39edd34fd34dad40202ec3b0b6eef"} Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.384651 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.459358 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-svc\") pod \"f38c0efe-db9f-4afc-8693-0743c558d74f\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.459432 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-swift-storage-0\") pod \"f38c0efe-db9f-4afc-8693-0743c558d74f\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.459549 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fvsq\" (UniqueName: \"kubernetes.io/projected/f38c0efe-db9f-4afc-8693-0743c558d74f-kube-api-access-6fvsq\") pod \"f38c0efe-db9f-4afc-8693-0743c558d74f\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.459609 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-sb\") pod \"f38c0efe-db9f-4afc-8693-0743c558d74f\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.459737 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-config\") pod \"f38c0efe-db9f-4afc-8693-0743c558d74f\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.459914 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-nb\") pod \"f38c0efe-db9f-4afc-8693-0743c558d74f\" (UID: \"f38c0efe-db9f-4afc-8693-0743c558d74f\") " Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.468486 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38c0efe-db9f-4afc-8693-0743c558d74f-kube-api-access-6fvsq" (OuterVolumeSpecName: "kube-api-access-6fvsq") pod "f38c0efe-db9f-4afc-8693-0743c558d74f" (UID: "f38c0efe-db9f-4afc-8693-0743c558d74f"). InnerVolumeSpecName "kube-api-access-6fvsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.517610 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f38c0efe-db9f-4afc-8693-0743c558d74f" (UID: "f38c0efe-db9f-4afc-8693-0743c558d74f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.519784 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f38c0efe-db9f-4afc-8693-0743c558d74f" (UID: "f38c0efe-db9f-4afc-8693-0743c558d74f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.531565 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-config" (OuterVolumeSpecName: "config") pod "f38c0efe-db9f-4afc-8693-0743c558d74f" (UID: "f38c0efe-db9f-4afc-8693-0743c558d74f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.533663 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f38c0efe-db9f-4afc-8693-0743c558d74f" (UID: "f38c0efe-db9f-4afc-8693-0743c558d74f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.542770 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f38c0efe-db9f-4afc-8693-0743c558d74f" (UID: "f38c0efe-db9f-4afc-8693-0743c558d74f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.562302 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.562526 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.562587 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fvsq\" (UniqueName: \"kubernetes.io/projected/f38c0efe-db9f-4afc-8693-0743c558d74f-kube-api-access-6fvsq\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.562640 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.562693 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:33 crc kubenswrapper[4778]: I0312 13:35:33.562780 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38c0efe-db9f-4afc-8693-0743c558d74f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.384384 4778 generic.go:334] "Generic (PLEG): container finished" podID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerID="f6f2930d9f70388763ddc5deccd561746f3634cc538e9cb6c56ef8628fd4e069" exitCode=0 Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.384497 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.384532 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerDied","Data":"f6f2930d9f70388763ddc5deccd561746f3634cc538e9cb6c56ef8628fd4e069"} Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.421543 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-xlfr7"] Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.430948 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-xlfr7"] Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.672118 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680276 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-combined-ca-bundle\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680355 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-run-httpd\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680392 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-config-data\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680412 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-ceilometer-tls-certs\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680472 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-sg-core-conf-yaml\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680505 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-log-httpd\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680536 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-scripts\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.680560 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fjcm\" (UniqueName: \"kubernetes.io/projected/bbcdf243-9822-4089-9cae-4a46417b6dc0-kube-api-access-2fjcm\") pod \"bbcdf243-9822-4089-9cae-4a46417b6dc0\" (UID: \"bbcdf243-9822-4089-9cae-4a46417b6dc0\") " Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.681142 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.681862 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.685463 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-scripts" (OuterVolumeSpecName: "scripts") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.685559 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbcdf243-9822-4089-9cae-4a46417b6dc0-kube-api-access-2fjcm" (OuterVolumeSpecName: "kube-api-access-2fjcm") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "kube-api-access-2fjcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.737381 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.744472 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.784994 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fjcm\" (UniqueName: \"kubernetes.io/projected/bbcdf243-9822-4089-9cae-4a46417b6dc0-kube-api-access-2fjcm\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.785032 4778 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.785045 4778 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.785055 4778 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.785066 4778 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbcdf243-9822-4089-9cae-4a46417b6dc0-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.785077 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.786784 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.803393 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-config-data" (OuterVolumeSpecName: "config-data") pod "bbcdf243-9822-4089-9cae-4a46417b6dc0" (UID: "bbcdf243-9822-4089-9cae-4a46417b6dc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.886681 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:34 crc kubenswrapper[4778]: I0312 13:35:34.886713 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbcdf243-9822-4089-9cae-4a46417b6dc0-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.396829 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbcdf243-9822-4089-9cae-4a46417b6dc0","Type":"ContainerDied","Data":"e349198afaff0969683d3154a99e49c5908b20bc0714e59a3832484e545b97dc"} Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.396885 4778 scope.go:117] "RemoveContainer" containerID="e8d17472cef396ced990a10ecac98a4762149d480cd7b1355d84ce3ecdbcf8ad" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.397019 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.440652 4778 scope.go:117] "RemoveContainer" containerID="8683a8e7549e2bde381c989208b511414ec56e8f866bf125984b6c4530f4d727" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.467820 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.474376 4778 scope.go:117] "RemoveContainer" containerID="7114be6621c79d0d604f29d2e6499dffdde39edd34fd34dad40202ec3b0b6eef" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.481978 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.526726 4778 scope.go:117] "RemoveContainer" containerID="f6f2930d9f70388763ddc5deccd561746f3634cc538e9cb6c56ef8628fd4e069" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.539319 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:35 crc kubenswrapper[4778]: E0312 13:35:35.540324 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-central-agent" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.540539 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-central-agent" Mar 12 13:35:35 crc kubenswrapper[4778]: E0312 13:35:35.540568 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="dnsmasq-dns" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.540580 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="dnsmasq-dns" Mar 12 13:35:35 crc kubenswrapper[4778]: E0312 13:35:35.540594 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="sg-core" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.540603 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="sg-core" Mar 12 13:35:35 crc kubenswrapper[4778]: E0312 13:35:35.540622 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-notification-agent" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.540630 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-notification-agent" Mar 12 13:35:35 crc kubenswrapper[4778]: E0312 13:35:35.540785 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="init" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.540846 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="init" Mar 12 13:35:35 crc kubenswrapper[4778]: E0312 13:35:35.540919 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="proxy-httpd" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.540966 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="proxy-httpd" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.541307 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-central-agent" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.541338 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="proxy-httpd" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.541380 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="ceilometer-notification-agent" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.541391 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="dnsmasq-dns" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.541407 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" containerName="sg-core" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.545007 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.548425 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.548759 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.548721 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.553389 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.724257 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.724645 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-scripts\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.724688 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f1d0355-a73a-4a93-94fb-b439436cf1b1-run-httpd\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.724914 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvzzb\" (UniqueName: \"kubernetes.io/projected/9f1d0355-a73a-4a93-94fb-b439436cf1b1-kube-api-access-tvzzb\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.725068 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-config-data\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.725112 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f1d0355-a73a-4a93-94fb-b439436cf1b1-log-httpd\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.725139 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.725205 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.826897 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f1d0355-a73a-4a93-94fb-b439436cf1b1-run-httpd\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827021 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvzzb\" (UniqueName: \"kubernetes.io/projected/9f1d0355-a73a-4a93-94fb-b439436cf1b1-kube-api-access-tvzzb\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827097 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-config-data\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827132 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f1d0355-a73a-4a93-94fb-b439436cf1b1-log-httpd\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827157 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827203 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827262 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827306 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-scripts\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827368 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f1d0355-a73a-4a93-94fb-b439436cf1b1-run-httpd\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.827731 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f1d0355-a73a-4a93-94fb-b439436cf1b1-log-httpd\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.833983 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.834275 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.834377 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-scripts\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.834966 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-config-data\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.842998 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1d0355-a73a-4a93-94fb-b439436cf1b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.848570 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvzzb\" (UniqueName: \"kubernetes.io/projected/9f1d0355-a73a-4a93-94fb-b439436cf1b1-kube-api-access-tvzzb\") pod \"ceilometer-0\" (UID: \"9f1d0355-a73a-4a93-94fb-b439436cf1b1\") " pod="openstack/ceilometer-0" Mar 12 13:35:35 crc kubenswrapper[4778]: I0312 13:35:35.864787 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 13:35:36 crc kubenswrapper[4778]: I0312 13:35:36.270739 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbcdf243-9822-4089-9cae-4a46417b6dc0" path="/var/lib/kubelet/pods/bbcdf243-9822-4089-9cae-4a46417b6dc0/volumes" Mar 12 13:35:36 crc kubenswrapper[4778]: I0312 13:35:36.272407 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" path="/var/lib/kubelet/pods/f38c0efe-db9f-4afc-8693-0743c558d74f/volumes" Mar 12 13:35:36 crc kubenswrapper[4778]: I0312 13:35:36.387346 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 13:35:36 crc kubenswrapper[4778]: I0312 13:35:36.429145 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f1d0355-a73a-4a93-94fb-b439436cf1b1","Type":"ContainerStarted","Data":"b98e3b9d38fb96804c09915afce6d36b14f7bb99c6ece9b42a84918c6e3c063e"} Mar 12 13:35:37 crc kubenswrapper[4778]: I0312 13:35:37.439465 4778 generic.go:334] "Generic (PLEG): container finished" podID="eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" containerID="a3547232ddc46df5ded5cc24fff2ec3e7c8bb4fb4c52277d66e27c319ec41995" exitCode=0 Mar 12 13:35:37 crc kubenswrapper[4778]: I0312 13:35:37.439550 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9xw6b" event={"ID":"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3","Type":"ContainerDied","Data":"a3547232ddc46df5ded5cc24fff2ec3e7c8bb4fb4c52277d66e27c319ec41995"} Mar 12 13:35:37 crc kubenswrapper[4778]: I0312 13:35:37.442157 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f1d0355-a73a-4a93-94fb-b439436cf1b1","Type":"ContainerStarted","Data":"86f3f82c78baa256cc0c6678d7d29c169da5465adc45d999aa4364ff7af57e50"} Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.363158 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-xlfr7" podUID="f38c0efe-db9f-4afc-8693-0743c558d74f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: i/o timeout" Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.452425 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f1d0355-a73a-4a93-94fb-b439436cf1b1","Type":"ContainerStarted","Data":"93ca9bbc7604492440314a22e964e6886cf6aa0937e7a02b01d0270505ad8f48"} Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.804856 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.994326 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-config-data\") pod \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.994437 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsjt8\" (UniqueName: \"kubernetes.io/projected/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-kube-api-access-zsjt8\") pod \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.994480 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-scripts\") pod \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.994500 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-combined-ca-bundle\") pod \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\" (UID: \"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3\") " Mar 12 13:35:38 crc kubenswrapper[4778]: I0312 13:35:38.998293 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-scripts" (OuterVolumeSpecName: "scripts") pod "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" (UID: "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.010714 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-kube-api-access-zsjt8" (OuterVolumeSpecName: "kube-api-access-zsjt8") pod "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" (UID: "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3"). InnerVolumeSpecName "kube-api-access-zsjt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.037364 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-config-data" (OuterVolumeSpecName: "config-data") pod "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" (UID: "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.046378 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" (UID: "eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.096540 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.096587 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsjt8\" (UniqueName: \"kubernetes.io/projected/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-kube-api-access-zsjt8\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.096598 4778 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.096609 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.462223 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9xw6b" event={"ID":"eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3","Type":"ContainerDied","Data":"fa86e251def50fe26c7890455a492370653fec6579cf29dd2f2d83fb340958c7"} Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.462569 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa86e251def50fe26c7890455a492370653fec6579cf29dd2f2d83fb340958c7" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.462383 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9xw6b" Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.465467 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f1d0355-a73a-4a93-94fb-b439436cf1b1","Type":"ContainerStarted","Data":"9f3f73eb3a2128645ff726985f9081f7cca012825c849d7c5d88697344bd1635"} Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.643833 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.644424 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-log" containerID="cri-o://7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397" gracePeriod=30 Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.645045 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-api" containerID="cri-o://eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11" gracePeriod=30 Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.664491 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.664765 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" containerName="nova-scheduler-scheduler" containerID="cri-o://dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" gracePeriod=30 Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.719856 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.720276 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-log" containerID="cri-o://ca24f0adae0376e480d75f053859a82c49878f39def1a1831162119084f0dc4d" gracePeriod=30 Mar 12 13:35:39 crc kubenswrapper[4778]: I0312 13:35:39.720494 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-metadata" containerID="cri-o://18451788e6f6468b69f6150e59f0635d08ad6db357c610ae673d149c136dfeb2" gracePeriod=30 Mar 12 13:35:41 crc kubenswrapper[4778]: E0312 13:35:41.780436 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:35:41 crc kubenswrapper[4778]: E0312 13:35:41.833374 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:35:41 crc kubenswrapper[4778]: E0312 13:35:41.883458 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:35:41 crc kubenswrapper[4778]: E0312 13:35:41.883535 4778 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" containerName="nova-scheduler-scheduler" Mar 12 13:35:41 crc kubenswrapper[4778]: I0312 13:35:41.891444 4778 generic.go:334] "Generic (PLEG): container finished" podID="c2700355-e048-4458-b430-8d149a08d624" containerID="ca24f0adae0376e480d75f053859a82c49878f39def1a1831162119084f0dc4d" exitCode=143 Mar 12 13:35:41 crc kubenswrapper[4778]: I0312 13:35:41.891538 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2700355-e048-4458-b430-8d149a08d624","Type":"ContainerDied","Data":"ca24f0adae0376e480d75f053859a82c49878f39def1a1831162119084f0dc4d"} Mar 12 13:35:41 crc kubenswrapper[4778]: I0312 13:35:41.907696 4778 generic.go:334] "Generic (PLEG): container finished" podID="a186b68c-e472-4507-abc7-0b90ca321ded" containerID="7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397" exitCode=143 Mar 12 13:35:41 crc kubenswrapper[4778]: I0312 13:35:41.907817 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a186b68c-e472-4507-abc7-0b90ca321ded","Type":"ContainerDied","Data":"7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397"} Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.599563 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.680074 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-combined-ca-bundle\") pod \"a186b68c-e472-4507-abc7-0b90ca321ded\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.680223 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-public-tls-certs\") pod \"a186b68c-e472-4507-abc7-0b90ca321ded\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.680270 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27cs2\" (UniqueName: \"kubernetes.io/projected/a186b68c-e472-4507-abc7-0b90ca321ded-kube-api-access-27cs2\") pod \"a186b68c-e472-4507-abc7-0b90ca321ded\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.680311 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-config-data\") pod \"a186b68c-e472-4507-abc7-0b90ca321ded\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.680389 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-internal-tls-certs\") pod \"a186b68c-e472-4507-abc7-0b90ca321ded\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.680412 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a186b68c-e472-4507-abc7-0b90ca321ded-logs\") pod \"a186b68c-e472-4507-abc7-0b90ca321ded\" (UID: \"a186b68c-e472-4507-abc7-0b90ca321ded\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.681155 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a186b68c-e472-4507-abc7-0b90ca321ded-logs" (OuterVolumeSpecName: "logs") pod "a186b68c-e472-4507-abc7-0b90ca321ded" (UID: "a186b68c-e472-4507-abc7-0b90ca321ded"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.687244 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a186b68c-e472-4507-abc7-0b90ca321ded-kube-api-access-27cs2" (OuterVolumeSpecName: "kube-api-access-27cs2") pod "a186b68c-e472-4507-abc7-0b90ca321ded" (UID: "a186b68c-e472-4507-abc7-0b90ca321ded"). InnerVolumeSpecName "kube-api-access-27cs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.752489 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a186b68c-e472-4507-abc7-0b90ca321ded" (UID: "a186b68c-e472-4507-abc7-0b90ca321ded"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.754878 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-config-data" (OuterVolumeSpecName: "config-data") pod "a186b68c-e472-4507-abc7-0b90ca321ded" (UID: "a186b68c-e472-4507-abc7-0b90ca321ded"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.758001 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a186b68c-e472-4507-abc7-0b90ca321ded" (UID: "a186b68c-e472-4507-abc7-0b90ca321ded"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.782775 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a186b68c-e472-4507-abc7-0b90ca321ded-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.782817 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.782828 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.782837 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27cs2\" (UniqueName: \"kubernetes.io/projected/a186b68c-e472-4507-abc7-0b90ca321ded-kube-api-access-27cs2\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.782845 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.785558 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a186b68c-e472-4507-abc7-0b90ca321ded" (UID: "a186b68c-e472-4507-abc7-0b90ca321ded"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.884767 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a186b68c-e472-4507-abc7-0b90ca321ded-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.904906 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.921030 4778 generic.go:334] "Generic (PLEG): container finished" podID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" exitCode=0 Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.921090 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53","Type":"ContainerDied","Data":"dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a"} Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.921115 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53","Type":"ContainerDied","Data":"65fd579a6354c4ba6b71d144e33926dc6f3bb53ede6b497f49fd5c5be99e7ee4"} Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.921133 4778 scope.go:117] "RemoveContainer" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.921298 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.935963 4778 generic.go:334] "Generic (PLEG): container finished" podID="a186b68c-e472-4507-abc7-0b90ca321ded" containerID="eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11" exitCode=0 Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.936040 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a186b68c-e472-4507-abc7-0b90ca321ded","Type":"ContainerDied","Data":"eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11"} Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.936074 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a186b68c-e472-4507-abc7-0b90ca321ded","Type":"ContainerDied","Data":"f5dd18bc2fa3f41fb9117d4f1e4c37d3b6b9987574f509d03e4a076b1981eb8c"} Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.936209 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.946852 4778 scope.go:117] "RemoveContainer" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" Mar 12 13:35:42 crc kubenswrapper[4778]: E0312 13:35:42.948008 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a\": container with ID starting with dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a not found: ID does not exist" containerID="dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.948065 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a"} err="failed to get container status \"dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a\": rpc error: code = NotFound desc = could not find container \"dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a\": container with ID starting with dae634b315afa4ad3533dd1e5963155a0f891be96e620c50199777eae097db0a not found: ID does not exist" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.948099 4778 scope.go:117] "RemoveContainer" containerID="eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.981278 4778 scope.go:117] "RemoveContainer" containerID="7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397" Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.986414 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-config-data\") pod \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.986586 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-combined-ca-bundle\") pod \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " Mar 12 13:35:42 crc kubenswrapper[4778]: I0312 13:35:42.986796 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xshlg\" (UniqueName: \"kubernetes.io/projected/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-kube-api-access-xshlg\") pod \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\" (UID: \"5bc6f909-0ff5-4f18-a480-fd8e6cda5e53\") " Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.001240 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-kube-api-access-xshlg" (OuterVolumeSpecName: "kube-api-access-xshlg") pod "5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" (UID: "5bc6f909-0ff5-4f18-a480-fd8e6cda5e53"). InnerVolumeSpecName "kube-api-access-xshlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.039413 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" (UID: "5bc6f909-0ff5-4f18-a480-fd8e6cda5e53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.056632 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-config-data" (OuterVolumeSpecName: "config-data") pod "5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" (UID: "5bc6f909-0ff5-4f18-a480-fd8e6cda5e53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.106517 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.106837 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.106915 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xshlg\" (UniqueName: \"kubernetes.io/projected/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53-kube-api-access-xshlg\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.236514 4778 scope.go:117] "RemoveContainer" containerID="eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11" Mar 12 13:35:43 crc kubenswrapper[4778]: E0312 13:35:43.237101 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11\": container with ID starting with eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11 not found: ID does not exist" containerID="eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.237130 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11"} err="failed to get container status \"eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11\": rpc error: code = NotFound desc = could not find container \"eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11\": container with ID starting with eaf94cdd79eea972e02bd1682954aa51e1069e729c7b0ae6e70982f28a03bc11 not found: ID does not exist" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.237166 4778 scope.go:117] "RemoveContainer" containerID="7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397" Mar 12 13:35:43 crc kubenswrapper[4778]: E0312 13:35:43.237474 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397\": container with ID starting with 7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397 not found: ID does not exist" containerID="7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.237525 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397"} err="failed to get container status \"7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397\": rpc error: code = NotFound desc = could not find container \"7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397\": container with ID starting with 7a2feca3aab730eaaa00a7eae47b95ddd0e61bb831e7003c0a078d8f2460d397 not found: ID does not exist" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.245597 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.256641 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.281710 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: E0312 13:35:43.282281 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" containerName="nova-manage" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282298 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" containerName="nova-manage" Mar 12 13:35:43 crc kubenswrapper[4778]: E0312 13:35:43.282329 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-api" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282340 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-api" Mar 12 13:35:43 crc kubenswrapper[4778]: E0312 13:35:43.282364 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" containerName="nova-scheduler-scheduler" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282372 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" containerName="nova-scheduler-scheduler" Mar 12 13:35:43 crc kubenswrapper[4778]: E0312 13:35:43.282386 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-log" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282394 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-log" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282627 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-log" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282642 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" containerName="nova-api-api" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282664 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" containerName="nova-manage" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.282678 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" containerName="nova-scheduler-scheduler" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.284248 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.289243 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.289541 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.289672 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.335498 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.357895 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.372106 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.387559 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.393559 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.395841 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.415783 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.425443 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-public-tls-certs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.425608 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d70066-6a42-4486-a487-e27b3ab3a61b-logs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.425687 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.427090 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-config-data\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.427183 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rctsf\" (UniqueName: \"kubernetes.io/projected/37d70066-6a42-4486-a487-e27b3ab3a61b-kube-api-access-rctsf\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.427246 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.528985 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530114 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-config-data\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530165 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530251 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rctsf\" (UniqueName: \"kubernetes.io/projected/37d70066-6a42-4486-a487-e27b3ab3a61b-kube-api-access-rctsf\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530314 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530489 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxrx6\" (UniqueName: \"kubernetes.io/projected/9c07bd9a-becb-4422-a881-5de27a8e8e56-kube-api-access-nxrx6\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530750 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-public-tls-certs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.530863 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-config-data\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.531067 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d70066-6a42-4486-a487-e27b3ab3a61b-logs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.531837 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d70066-6a42-4486-a487-e27b3ab3a61b-logs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.537771 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.539077 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-public-tls-certs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.540299 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.550979 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-config-data\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.558810 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rctsf\" (UniqueName: \"kubernetes.io/projected/37d70066-6a42-4486-a487-e27b3ab3a61b-kube-api-access-rctsf\") pod \"nova-api-0\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.612598 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.632572 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.632659 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxrx6\" (UniqueName: \"kubernetes.io/projected/9c07bd9a-becb-4422-a881-5de27a8e8e56-kube-api-access-nxrx6\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.632745 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-config-data\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.637363 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-config-data\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.639043 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.656218 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxrx6\" (UniqueName: \"kubernetes.io/projected/9c07bd9a-becb-4422-a881-5de27a8e8e56-kube-api-access-nxrx6\") pod \"nova-scheduler-0\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.717736 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.952718 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f1d0355-a73a-4a93-94fb-b439436cf1b1","Type":"ContainerStarted","Data":"4bdb99a1b06b9bc7b476cd5b6e145e69cc5597c366ddaf40ed2afb2b51af7963"} Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.954680 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 13:35:43 crc kubenswrapper[4778]: I0312 13:35:43.985817 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.324860715 podStartE2EDuration="8.985793387s" podCreationTimestamp="2026-03-12 13:35:35 +0000 UTC" firstStartedPulling="2026-03-12 13:35:36.415492889 +0000 UTC m=+1554.864188285" lastFinishedPulling="2026-03-12 13:35:43.076425561 +0000 UTC m=+1561.525120957" observedRunningTime="2026-03-12 13:35:43.980908228 +0000 UTC m=+1562.429603634" watchObservedRunningTime="2026-03-12 13:35:43.985793387 +0000 UTC m=+1562.434488783" Mar 12 13:35:44 crc kubenswrapper[4778]: I0312 13:35:44.096062 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:35:44 crc kubenswrapper[4778]: W0312 13:35:44.102828 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c07bd9a_becb_4422_a881_5de27a8e8e56.slice/crio-7c66f4ab481a2ffe8aa1f978637b76237ffd5a0742d58abdd29d2a665dd400d8 WatchSource:0}: Error finding container 7c66f4ab481a2ffe8aa1f978637b76237ffd5a0742d58abdd29d2a665dd400d8: Status 404 returned error can't find the container with id 7c66f4ab481a2ffe8aa1f978637b76237ffd5a0742d58abdd29d2a665dd400d8 Mar 12 13:35:44 crc kubenswrapper[4778]: I0312 13:35:44.130287 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:35:44 crc kubenswrapper[4778]: I0312 13:35:44.270754 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc6f909-0ff5-4f18-a480-fd8e6cda5e53" path="/var/lib/kubelet/pods/5bc6f909-0ff5-4f18-a480-fd8e6cda5e53/volumes" Mar 12 13:35:44 crc kubenswrapper[4778]: I0312 13:35:44.271909 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a186b68c-e472-4507-abc7-0b90ca321ded" path="/var/lib/kubelet/pods/a186b68c-e472-4507-abc7-0b90ca321ded/volumes" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.005424 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c07bd9a-becb-4422-a881-5de27a8e8e56","Type":"ContainerStarted","Data":"76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549"} Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.005739 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c07bd9a-becb-4422-a881-5de27a8e8e56","Type":"ContainerStarted","Data":"7c66f4ab481a2ffe8aa1f978637b76237ffd5a0742d58abdd29d2a665dd400d8"} Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.023517 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37d70066-6a42-4486-a487-e27b3ab3a61b","Type":"ContainerStarted","Data":"6fb0ea63ecde6cfac6694eb778a0f0043874e52ed36561d1c373be870defe193"} Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.023575 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37d70066-6a42-4486-a487-e27b3ab3a61b","Type":"ContainerStarted","Data":"2b4aa266ad205b7e0a6d8899547a75fe40c64017eb43d076bced61bb7cc36c19"} Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.023592 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37d70066-6a42-4486-a487-e27b3ab3a61b","Type":"ContainerStarted","Data":"c412c8241a59192093777ad55c60d09316d57b1f207c8116b8342fac0e609d85"} Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.062567 4778 generic.go:334] "Generic (PLEG): container finished" podID="c2700355-e048-4458-b430-8d149a08d624" containerID="18451788e6f6468b69f6150e59f0635d08ad6db357c610ae673d149c136dfeb2" exitCode=0 Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.064103 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2700355-e048-4458-b430-8d149a08d624","Type":"ContainerDied","Data":"18451788e6f6468b69f6150e59f0635d08ad6db357c610ae673d149c136dfeb2"} Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.119535 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.119507566 podStartE2EDuration="2.119507566s" podCreationTimestamp="2026-03-12 13:35:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:45.074623843 +0000 UTC m=+1563.523319239" watchObservedRunningTime="2026-03-12 13:35:45.119507566 +0000 UTC m=+1563.568202962" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.119867 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.119856536 podStartE2EDuration="2.119856536s" podCreationTimestamp="2026-03-12 13:35:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:45.107376042 +0000 UTC m=+1563.556071438" watchObservedRunningTime="2026-03-12 13:35:45.119856536 +0000 UTC m=+1563.568551932" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.437094 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.607265 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-config-data\") pod \"c2700355-e048-4458-b430-8d149a08d624\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.607715 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbjl2\" (UniqueName: \"kubernetes.io/projected/c2700355-e048-4458-b430-8d149a08d624-kube-api-access-jbjl2\") pod \"c2700355-e048-4458-b430-8d149a08d624\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.608090 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-nova-metadata-tls-certs\") pod \"c2700355-e048-4458-b430-8d149a08d624\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.608284 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-combined-ca-bundle\") pod \"c2700355-e048-4458-b430-8d149a08d624\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.608375 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2700355-e048-4458-b430-8d149a08d624-logs\") pod \"c2700355-e048-4458-b430-8d149a08d624\" (UID: \"c2700355-e048-4458-b430-8d149a08d624\") " Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.609428 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2700355-e048-4458-b430-8d149a08d624-logs" (OuterVolumeSpecName: "logs") pod "c2700355-e048-4458-b430-8d149a08d624" (UID: "c2700355-e048-4458-b430-8d149a08d624"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.614374 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2700355-e048-4458-b430-8d149a08d624-kube-api-access-jbjl2" (OuterVolumeSpecName: "kube-api-access-jbjl2") pod "c2700355-e048-4458-b430-8d149a08d624" (UID: "c2700355-e048-4458-b430-8d149a08d624"). InnerVolumeSpecName "kube-api-access-jbjl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.652438 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-config-data" (OuterVolumeSpecName: "config-data") pod "c2700355-e048-4458-b430-8d149a08d624" (UID: "c2700355-e048-4458-b430-8d149a08d624"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.663101 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2700355-e048-4458-b430-8d149a08d624" (UID: "c2700355-e048-4458-b430-8d149a08d624"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.683375 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c2700355-e048-4458-b430-8d149a08d624" (UID: "c2700355-e048-4458-b430-8d149a08d624"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.710836 4778 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.710873 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.710882 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2700355-e048-4458-b430-8d149a08d624-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.710893 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2700355-e048-4458-b430-8d149a08d624-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:45 crc kubenswrapper[4778]: I0312 13:35:45.710901 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbjl2\" (UniqueName: \"kubernetes.io/projected/c2700355-e048-4458-b430-8d149a08d624-kube-api-access-jbjl2\") on node \"crc\" DevicePath \"\"" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.074984 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.077389 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2700355-e048-4458-b430-8d149a08d624","Type":"ContainerDied","Data":"4cdf23596db0a1e92716b41df3f9c56dd37f21ba73a9653782acae39684ee3dd"} Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.077463 4778 scope.go:117] "RemoveContainer" containerID="18451788e6f6468b69f6150e59f0635d08ad6db357c610ae673d149c136dfeb2" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.114637 4778 scope.go:117] "RemoveContainer" containerID="ca24f0adae0376e480d75f053859a82c49878f39def1a1831162119084f0dc4d" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.132515 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.146270 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.157435 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:35:46 crc kubenswrapper[4778]: E0312 13:35:46.157914 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-metadata" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.157943 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-metadata" Mar 12 13:35:46 crc kubenswrapper[4778]: E0312 13:35:46.157974 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-log" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.157983 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-log" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.158263 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-metadata" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.158296 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2700355-e048-4458-b430-8d149a08d624" containerName="nova-metadata-log" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.159599 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.164316 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.164499 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.191418 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.265884 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2700355-e048-4458-b430-8d149a08d624" path="/var/lib/kubelet/pods/c2700355-e048-4458-b430-8d149a08d624/volumes" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.322127 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5dsf\" (UniqueName: \"kubernetes.io/projected/e5f2bac2-0571-44d8-ba4e-c006600506a5-kube-api-access-d5dsf\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.322187 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.322350 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2bac2-0571-44d8-ba4e-c006600506a5-logs\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.322369 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.322772 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-config-data\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.424518 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-config-data\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.424579 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5dsf\" (UniqueName: \"kubernetes.io/projected/e5f2bac2-0571-44d8-ba4e-c006600506a5-kube-api-access-d5dsf\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.424605 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.424678 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2bac2-0571-44d8-ba4e-c006600506a5-logs\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.424696 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.425784 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2bac2-0571-44d8-ba4e-c006600506a5-logs\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.432746 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.433629 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.440282 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-config-data\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.443922 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5dsf\" (UniqueName: \"kubernetes.io/projected/e5f2bac2-0571-44d8-ba4e-c006600506a5-kube-api-access-d5dsf\") pod \"nova-metadata-0\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.484480 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:35:46 crc kubenswrapper[4778]: I0312 13:35:46.929579 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:35:47 crc kubenswrapper[4778]: I0312 13:35:47.093857 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5f2bac2-0571-44d8-ba4e-c006600506a5","Type":"ContainerStarted","Data":"63ddaf1ae0c88152b73f19a7bbf611a71857e4349ca72eb4c73d9d4e815e1b3c"} Mar 12 13:35:48 crc kubenswrapper[4778]: I0312 13:35:48.104896 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5f2bac2-0571-44d8-ba4e-c006600506a5","Type":"ContainerStarted","Data":"1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed"} Mar 12 13:35:48 crc kubenswrapper[4778]: I0312 13:35:48.105060 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5f2bac2-0571-44d8-ba4e-c006600506a5","Type":"ContainerStarted","Data":"3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c"} Mar 12 13:35:48 crc kubenswrapper[4778]: I0312 13:35:48.137894 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.137869864 podStartE2EDuration="2.137869864s" podCreationTimestamp="2026-03-12 13:35:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:35:48.129789655 +0000 UTC m=+1566.578485061" watchObservedRunningTime="2026-03-12 13:35:48.137869864 +0000 UTC m=+1566.586565260" Mar 12 13:35:48 crc kubenswrapper[4778]: I0312 13:35:48.719357 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 13:35:53 crc kubenswrapper[4778]: I0312 13:35:53.613342 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:35:53 crc kubenswrapper[4778]: I0312 13:35:53.613968 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:35:53 crc kubenswrapper[4778]: I0312 13:35:53.718777 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 12 13:35:53 crc kubenswrapper[4778]: I0312 13:35:53.746174 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 12 13:35:54 crc kubenswrapper[4778]: I0312 13:35:54.202708 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 12 13:35:54 crc kubenswrapper[4778]: I0312 13:35:54.630456 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:54 crc kubenswrapper[4778]: I0312 13:35:54.630479 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:55 crc kubenswrapper[4778]: I0312 13:35:55.448377 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e1488e83-3a44-41ad-aa96-de09b662c16e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.198:3000/\": dial tcp 10.217.0.198:3000: i/o timeout" Mar 12 13:35:56 crc kubenswrapper[4778]: I0312 13:35:56.485270 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 13:35:56 crc kubenswrapper[4778]: I0312 13:35:56.485595 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 13:35:57 crc kubenswrapper[4778]: I0312 13:35:57.497377 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:57 crc kubenswrapper[4778]: I0312 13:35:57.497381 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:35:58 crc kubenswrapper[4778]: I0312 13:35:58.557913 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:35:58 crc kubenswrapper[4778]: I0312 13:35:58.557961 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.157016 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555376-2bdpv"] Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.158690 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.162255 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.162560 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.163867 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.168434 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555376-2bdpv"] Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.309408 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvvbv\" (UniqueName: \"kubernetes.io/projected/1f74db3c-fec4-452d-bfd6-8db9f766e0bc-kube-api-access-mvvbv\") pod \"auto-csr-approver-29555376-2bdpv\" (UID: \"1f74db3c-fec4-452d-bfd6-8db9f766e0bc\") " pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.411590 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvvbv\" (UniqueName: \"kubernetes.io/projected/1f74db3c-fec4-452d-bfd6-8db9f766e0bc-kube-api-access-mvvbv\") pod \"auto-csr-approver-29555376-2bdpv\" (UID: \"1f74db3c-fec4-452d-bfd6-8db9f766e0bc\") " pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.456541 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvvbv\" (UniqueName: \"kubernetes.io/projected/1f74db3c-fec4-452d-bfd6-8db9f766e0bc-kube-api-access-mvvbv\") pod \"auto-csr-approver-29555376-2bdpv\" (UID: \"1f74db3c-fec4-452d-bfd6-8db9f766e0bc\") " pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.478825 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:00 crc kubenswrapper[4778]: I0312 13:36:00.928567 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555376-2bdpv"] Mar 12 13:36:01 crc kubenswrapper[4778]: I0312 13:36:01.239360 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" event={"ID":"1f74db3c-fec4-452d-bfd6-8db9f766e0bc","Type":"ContainerStarted","Data":"211e7ee34a5dd3127d5e72bf488445d32209f3b42800d63aa54c4f05c30abbd1"} Mar 12 13:36:01 crc kubenswrapper[4778]: I0312 13:36:01.613148 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 13:36:01 crc kubenswrapper[4778]: I0312 13:36:01.614252 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 13:36:03 crc kubenswrapper[4778]: I0312 13:36:03.262309 4778 generic.go:334] "Generic (PLEG): container finished" podID="1f74db3c-fec4-452d-bfd6-8db9f766e0bc" containerID="0e8b3287f4617d49763a5e13085485c7f1faa35a7b545d67c3db4b7ac7a3c06b" exitCode=0 Mar 12 13:36:03 crc kubenswrapper[4778]: I0312 13:36:03.263293 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" event={"ID":"1f74db3c-fec4-452d-bfd6-8db9f766e0bc","Type":"ContainerDied","Data":"0e8b3287f4617d49763a5e13085485c7f1faa35a7b545d67c3db4b7ac7a3c06b"} Mar 12 13:36:03 crc kubenswrapper[4778]: I0312 13:36:03.620501 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 13:36:03 crc kubenswrapper[4778]: I0312 13:36:03.621218 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 13:36:03 crc kubenswrapper[4778]: I0312 13:36:03.635301 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 13:36:03 crc kubenswrapper[4778]: I0312 13:36:03.635638 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 13:36:04 crc kubenswrapper[4778]: I0312 13:36:04.485571 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 13:36:04 crc kubenswrapper[4778]: I0312 13:36:04.485631 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 13:36:04 crc kubenswrapper[4778]: I0312 13:36:04.632921 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:04 crc kubenswrapper[4778]: I0312 13:36:04.814653 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvvbv\" (UniqueName: \"kubernetes.io/projected/1f74db3c-fec4-452d-bfd6-8db9f766e0bc-kube-api-access-mvvbv\") pod \"1f74db3c-fec4-452d-bfd6-8db9f766e0bc\" (UID: \"1f74db3c-fec4-452d-bfd6-8db9f766e0bc\") " Mar 12 13:36:04 crc kubenswrapper[4778]: I0312 13:36:04.861568 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f74db3c-fec4-452d-bfd6-8db9f766e0bc-kube-api-access-mvvbv" (OuterVolumeSpecName: "kube-api-access-mvvbv") pod "1f74db3c-fec4-452d-bfd6-8db9f766e0bc" (UID: "1f74db3c-fec4-452d-bfd6-8db9f766e0bc"). InnerVolumeSpecName "kube-api-access-mvvbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:04 crc kubenswrapper[4778]: I0312 13:36:04.916583 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvvbv\" (UniqueName: \"kubernetes.io/projected/1f74db3c-fec4-452d-bfd6-8db9f766e0bc-kube-api-access-mvvbv\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:05 crc kubenswrapper[4778]: I0312 13:36:05.289841 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" Mar 12 13:36:05 crc kubenswrapper[4778]: I0312 13:36:05.290046 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555376-2bdpv" event={"ID":"1f74db3c-fec4-452d-bfd6-8db9f766e0bc","Type":"ContainerDied","Data":"211e7ee34a5dd3127d5e72bf488445d32209f3b42800d63aa54c4f05c30abbd1"} Mar 12 13:36:05 crc kubenswrapper[4778]: I0312 13:36:05.290381 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="211e7ee34a5dd3127d5e72bf488445d32209f3b42800d63aa54c4f05c30abbd1" Mar 12 13:36:05 crc kubenswrapper[4778]: I0312 13:36:05.716054 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555370-6zrgd"] Mar 12 13:36:05 crc kubenswrapper[4778]: I0312 13:36:05.732496 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555370-6zrgd"] Mar 12 13:36:05 crc kubenswrapper[4778]: I0312 13:36:05.890160 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 12 13:36:06 crc kubenswrapper[4778]: I0312 13:36:06.262792 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c682acb-240b-44d4-a2be-0ea0cd913af1" path="/var/lib/kubelet/pods/1c682acb-240b-44d4-a2be-0ea0cd913af1/volumes" Mar 12 13:36:06 crc kubenswrapper[4778]: I0312 13:36:06.494543 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 12 13:36:06 crc kubenswrapper[4778]: I0312 13:36:06.496863 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 12 13:36:06 crc kubenswrapper[4778]: I0312 13:36:06.500867 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 12 13:36:07 crc kubenswrapper[4778]: I0312 13:36:07.312828 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 12 13:36:15 crc kubenswrapper[4778]: I0312 13:36:15.711863 4778 scope.go:117] "RemoveContainer" containerID="8328194fef169053b3f39722ffd3e2d940869363b5142050b8e768ed01fab0c0" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.764543 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-69b6dc4885-z4h9m"] Mar 12 13:36:23 crc kubenswrapper[4778]: E0312 13:36:23.766082 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f74db3c-fec4-452d-bfd6-8db9f766e0bc" containerName="oc" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.766109 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f74db3c-fec4-452d-bfd6-8db9f766e0bc" containerName="oc" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.766424 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f74db3c-fec4-452d-bfd6-8db9f766e0bc" containerName="oc" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.767948 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.786979 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-69b6dc4885-z4h9m"] Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812195 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-scripts\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812274 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-public-tls-certs\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812308 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-fernet-keys\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812335 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-combined-ca-bundle\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812413 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-credential-keys\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812473 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9h9j\" (UniqueName: \"kubernetes.io/projected/16dea17b-eaa4-4bbf-8895-c077b3e28d66-kube-api-access-d9h9j\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812946 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-internal-tls-certs\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.812994 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-config-data\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.914982 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-internal-tls-certs\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915034 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-config-data\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915070 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-scripts\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915104 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-public-tls-certs\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915130 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-fernet-keys\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915150 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-combined-ca-bundle\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915197 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-credential-keys\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.915244 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9h9j\" (UniqueName: \"kubernetes.io/projected/16dea17b-eaa4-4bbf-8895-c077b3e28d66-kube-api-access-d9h9j\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.921646 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-scripts\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.921749 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-fernet-keys\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.922100 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-credential-keys\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.922104 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-combined-ca-bundle\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.923219 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-public-tls-certs\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.925615 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-internal-tls-certs\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.930475 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dea17b-eaa4-4bbf-8895-c077b3e28d66-config-data\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:23 crc kubenswrapper[4778]: I0312 13:36:23.934438 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9h9j\" (UniqueName: \"kubernetes.io/projected/16dea17b-eaa4-4bbf-8895-c077b3e28d66-kube-api-access-d9h9j\") pod \"keystone-69b6dc4885-z4h9m\" (UID: \"16dea17b-eaa4-4bbf-8895-c077b3e28d66\") " pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.103077 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.122789 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-769c65dfd5-frvxx"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.132984 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.165138 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-769c65dfd5-frvxx"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.189969 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-566c4d5fc-zx97x"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.192351 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.222405 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kj5b\" (UniqueName: \"kubernetes.io/projected/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-kube-api-access-4kj5b\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.222769 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-combined-ca-bundle\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.222812 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-httpd-config\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.222838 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-ovndb-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.222890 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-internal-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.222986 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-ovndb-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223013 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-combined-ca-bundle\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223047 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqjss\" (UniqueName: \"kubernetes.io/projected/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-kube-api-access-nqjss\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223163 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-config\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223251 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-internal-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223284 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-public-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223316 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-config\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223354 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-httpd-config\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.223407 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-public-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.234135 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-566c4d5fc-zx97x"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325432 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-config\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325518 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-httpd-config\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325566 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-public-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325600 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kj5b\" (UniqueName: \"kubernetes.io/projected/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-kube-api-access-4kj5b\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325628 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-combined-ca-bundle\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325666 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-ovndb-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325693 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-httpd-config\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325736 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-internal-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325825 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-ovndb-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325852 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-combined-ca-bundle\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.325885 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqjss\" (UniqueName: \"kubernetes.io/projected/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-kube-api-access-nqjss\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.326033 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-config\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.326148 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-internal-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.326203 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-public-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.332703 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-config\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.334604 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-ovndb-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.342111 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-httpd-config\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.343850 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-public-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.343877 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-httpd-config\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.346518 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kj5b\" (UniqueName: \"kubernetes.io/projected/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-kube-api-access-4kj5b\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.346542 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-internal-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.347176 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-combined-ca-bundle\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.353466 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-ovndb-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.355493 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-public-tls-certs\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.361904 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqjss\" (UniqueName: \"kubernetes.io/projected/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-kube-api-access-nqjss\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.372923 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-internal-tls-certs\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.375072 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67d4b7-d8eb-40f4-b51d-62e92c6042c1-combined-ca-bundle\") pod \"neutron-566c4d5fc-zx97x\" (UID: \"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1\") " pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.400670 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-config\") pod \"neutron-769c65dfd5-frvxx\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.497512 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.497839 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" containerName="nova-cell0-conductor-conductor" containerID="cri-o://7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731" gracePeriod=30 Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.519969 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.520380 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9c07bd9a-becb-4422-a881-5de27a8e8e56" containerName="nova-scheduler-scheduler" containerID="cri-o://76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" gracePeriod=30 Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.569974 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.570271 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-log" containerID="cri-o://2b4aa266ad205b7e0a6d8899547a75fe40c64017eb43d076bced61bb7cc36c19" gracePeriod=30 Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.570410 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-api" containerID="cri-o://6fb0ea63ecde6cfac6694eb778a0f0043874e52ed36561d1c373be870defe193" gracePeriod=30 Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.595847 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:24 crc kubenswrapper[4778]: I0312 13:36:24.597470 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.607775 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.641978 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-69b6dc4885-z4h9m"] Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.734226 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.745377 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-log" containerID="cri-o://3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c" gracePeriod=30 Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.745858 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-metadata" containerID="cri-o://1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed" gracePeriod=30 Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.793323 4778 generic.go:334] "Generic (PLEG): container finished" podID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerID="2b4aa266ad205b7e0a6d8899547a75fe40c64017eb43d076bced61bb7cc36c19" exitCode=143 Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.793381 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37d70066-6a42-4486-a487-e27b3ab3a61b","Type":"ContainerDied","Data":"2b4aa266ad205b7e0a6d8899547a75fe40c64017eb43d076bced61bb7cc36c19"} Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.803336 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1"] Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.830591 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.892829 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1"] Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.928164 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.936458 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.940051 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-metadata-config-data" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.968459 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5z5t\" (UniqueName: \"kubernetes.io/projected/f0341d80-4327-4c9e-bc11-0cddbc6eab66-kube-api-access-t5z5t\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.978241 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-internal-tls-certs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.978350 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0341d80-4327-4c9e-bc11-0cddbc6eab66-logs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.978410 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-public-tls-certs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.978442 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-combined-ca-bundle\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.978679 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-config-data\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:25 crc kubenswrapper[4778]: I0312 13:36:25.985030 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.022198 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6796c46585-tk69s"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.081863 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-public-tls-certs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.081927 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a29f7b0-d851-4967-802b-91e301ce82f2-logs\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.081962 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-combined-ca-bundle\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082045 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-config-data\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082091 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-config-data\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082163 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5z5t\" (UniqueName: \"kubernetes.io/projected/f0341d80-4327-4c9e-bc11-0cddbc6eab66-kube-api-access-t5z5t\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082397 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-combined-ca-bundle\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082454 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-internal-tls-certs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082495 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4cqk\" (UniqueName: \"kubernetes.io/projected/5a29f7b0-d851-4967-802b-91e301ce82f2-kube-api-access-m4cqk\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.082547 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0341d80-4327-4c9e-bc11-0cddbc6eab66-logs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.084500 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0341d80-4327-4c9e-bc11-0cddbc6eab66-logs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.103020 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.112807 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-internal-tls-certs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.134564 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-config-data\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.134859 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-public-tls-certs\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.153779 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5z5t\" (UniqueName: \"kubernetes.io/projected/f0341d80-4327-4c9e-bc11-0cddbc6eab66-kube-api-access-t5z5t\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.188097 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0341d80-4327-4c9e-bc11-0cddbc6eab66-combined-ca-bundle\") pod \"nova-api-1\" (UID: \"f0341d80-4327-4c9e-bc11-0cddbc6eab66\") " pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.189426 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-combined-ca-bundle\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.192097 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-config\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.192323 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4cqk\" (UniqueName: \"kubernetes.io/projected/5a29f7b0-d851-4967-802b-91e301ce82f2-kube-api-access-m4cqk\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.193369 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a29f7b0-d851-4967-802b-91e301ce82f2-logs\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.193489 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-svc\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.202126 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a29f7b0-d851-4967-802b-91e301ce82f2-logs\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.224765 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc9pp\" (UniqueName: \"kubernetes.io/projected/7be5cd74-51aa-4be4-bee0-bcd4414e988c-kube-api-access-mc9pp\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.225010 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-config-data\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.225299 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-nb\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.225392 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-sb\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.225515 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-swift-storage-0\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.246067 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4cqk\" (UniqueName: \"kubernetes.io/projected/5a29f7b0-d851-4967-802b-91e301ce82f2-kube-api-access-m4cqk\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.246896 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.254434 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-config-data\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.292431 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6796c46585-tk69s"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.306040 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-combined-ca-bundle\") pod \"nova-cell1-metadata-0\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.330926 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-config\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.331371 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.333115 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-svc\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.334031 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc9pp\" (UniqueName: \"kubernetes.io/projected/7be5cd74-51aa-4be4-bee0-bcd4414e988c-kube-api-access-mc9pp\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.334520 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-nb\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.334573 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-sb\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.334621 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-swift-storage-0\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.336291 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-swift-storage-0\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.342801 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-nb\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.343409 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-sb\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.345019 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-config\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.373028 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-svc\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.376929 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc9pp\" (UniqueName: \"kubernetes.io/projected/7be5cd74-51aa-4be4-bee0-bcd4414e988c-kube-api-access-mc9pp\") pod \"dnsmasq-dns-6796c46585-tk69s\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.396237 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6796c46585-tk69s"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.396970 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.439484 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f59c7d6f9-7f6bj"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.441336 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.477123 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f59c7d6f9-7f6bj"] Mar 12 13:36:26 crc kubenswrapper[4778]: E0312 13:36:26.534036 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 12 13:36:26 crc kubenswrapper[4778]: E0312 13:36:26.537282 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.539494 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdn6k\" (UniqueName: \"kubernetes.io/projected/a677b5ba-f5d3-4310-ab6d-af0505e82a00-kube-api-access-tdn6k\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.539549 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-nb\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.539593 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-swift-storage-0\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.539628 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-sb\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.540049 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-config\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.540224 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-svc\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: E0312 13:36:26.547662 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 12 13:36:26 crc kubenswrapper[4778]: E0312 13:36:26.547750 4778 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" containerName="nova-cell0-conductor-conductor" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.647990 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdn6k\" (UniqueName: \"kubernetes.io/projected/a677b5ba-f5d3-4310-ab6d-af0505e82a00-kube-api-access-tdn6k\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.648059 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-nb\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.648119 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-swift-storage-0\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.648145 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-sb\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.648243 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-config\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.648298 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-svc\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.650718 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-nb\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.650831 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-sb\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.649442 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-svc\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.651031 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-config\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.651148 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-swift-storage-0\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.684885 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdn6k\" (UniqueName: \"kubernetes.io/projected/a677b5ba-f5d3-4310-ab6d-af0505e82a00-kube-api-access-tdn6k\") pod \"dnsmasq-dns-6f59c7d6f9-7f6bj\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.689322 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-769c65dfd5-frvxx"] Mar 12 13:36:26 crc kubenswrapper[4778]: W0312 13:36:26.703067 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e1d8894_7234_40d0_b42a_9d7ab1ce638a.slice/crio-e2d775202948449d32b3e6f8c66299f17943aeca0f3f57c7b82f6f8283ff7095 WatchSource:0}: Error finding container e2d775202948449d32b3e6f8c66299f17943aeca0f3f57c7b82f6f8283ff7095: Status 404 returned error can't find the container with id e2d775202948449d32b3e6f8c66299f17943aeca0f3f57c7b82f6f8283ff7095 Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.764603 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.792604 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-566c4d5fc-zx97x"] Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.836027 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-69b6dc4885-z4h9m" event={"ID":"16dea17b-eaa4-4bbf-8895-c077b3e28d66","Type":"ContainerStarted","Data":"6fcce949c75d6468b3b14633abe6749e40c68a0baec2473ee7c3eb866c1fae55"} Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.836408 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-69b6dc4885-z4h9m" event={"ID":"16dea17b-eaa4-4bbf-8895-c077b3e28d66","Type":"ContainerStarted","Data":"29a78f7251ac4b00bd393718f7a5b34295387492a9278080f3668886800c6546"} Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.839025 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.844260 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c4d5fc-zx97x" event={"ID":"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1","Type":"ContainerStarted","Data":"6a6275efc50e33030547560ccee86be3358a86873c23828c4bdef28e7387e6ec"} Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.854023 4778 generic.go:334] "Generic (PLEG): container finished" podID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerID="3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c" exitCode=143 Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.854145 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5f2bac2-0571-44d8-ba4e-c006600506a5","Type":"ContainerDied","Data":"3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c"} Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.856196 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" containerName="nova-cell1-conductor-conductor" containerID="cri-o://17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f" gracePeriod=30 Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.856814 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-frvxx" event={"ID":"2e1d8894-7234-40d0-b42a-9d7ab1ce638a","Type":"ContainerStarted","Data":"e2d775202948449d32b3e6f8c66299f17943aeca0f3f57c7b82f6f8283ff7095"} Mar 12 13:36:26 crc kubenswrapper[4778]: I0312 13:36:26.889925 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-69b6dc4885-z4h9m" podStartSLOduration=3.889903946 podStartE2EDuration="3.889903946s" podCreationTimestamp="2026-03-12 13:36:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:26.888546048 +0000 UTC m=+1605.337241454" watchObservedRunningTime="2026-03-12 13:36:26.889903946 +0000 UTC m=+1605.338599342" Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.011147 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1"] Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.386831 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6796c46585-tk69s"] Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.418856 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:27 crc kubenswrapper[4778]: W0312 13:36:27.481332 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7be5cd74_51aa_4be4_bee0_bcd4414e988c.slice/crio-9611fdd9f50898df9bc39508d9d61ff6e00eab70329522bacb24a65cad5f58f2 WatchSource:0}: Error finding container 9611fdd9f50898df9bc39508d9d61ff6e00eab70329522bacb24a65cad5f58f2: Status 404 returned error can't find the container with id 9611fdd9f50898df9bc39508d9d61ff6e00eab70329522bacb24a65cad5f58f2 Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.616419 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f59c7d6f9-7f6bj"] Mar 12 13:36:27 crc kubenswrapper[4778]: E0312 13:36:27.702471 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7733a48b_2bc4_4372_a222_37bb8ea04b6d.slice/crio-7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7733a48b_2bc4_4372_a222_37bb8ea04b6d.slice/crio-conmon-7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731.scope\": RecentStats: unable to find data in memory cache]" Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.939374 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.940963 4778 generic.go:334] "Generic (PLEG): container finished" podID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerID="6fb0ea63ecde6cfac6694eb778a0f0043874e52ed36561d1c373be870defe193" exitCode=0 Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.941091 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37d70066-6a42-4486-a487-e27b3ab3a61b","Type":"ContainerDied","Data":"6fb0ea63ecde6cfac6694eb778a0f0043874e52ed36561d1c373be870defe193"} Mar 12 13:36:27 crc kubenswrapper[4778]: I0312 13:36:27.974680 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6796c46585-tk69s" event={"ID":"7be5cd74-51aa-4be4-bee0-bcd4414e988c","Type":"ContainerStarted","Data":"9611fdd9f50898df9bc39508d9d61ff6e00eab70329522bacb24a65cad5f58f2"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.005907 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xht95\" (UniqueName: \"kubernetes.io/projected/7733a48b-2bc4-4372-a222-37bb8ea04b6d-kube-api-access-xht95\") pod \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.005998 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-combined-ca-bundle\") pod \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.006203 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-config-data\") pod \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\" (UID: \"7733a48b-2bc4-4372-a222-37bb8ea04b6d\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.024574 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7733a48b-2bc4-4372-a222-37bb8ea04b6d-kube-api-access-xht95" (OuterVolumeSpecName: "kube-api-access-xht95") pod "7733a48b-2bc4-4372-a222-37bb8ea04b6d" (UID: "7733a48b-2bc4-4372-a222-37bb8ea04b6d"). InnerVolumeSpecName "kube-api-access-xht95". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.052121 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"5a29f7b0-d851-4967-802b-91e301ce82f2","Type":"ContainerStarted","Data":"7686b13a63fb7303c82944b69b0f74a27bb498a86b8a8900db3aa379dab6a697"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.081236 4778 generic.go:334] "Generic (PLEG): container finished" podID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" containerID="7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731" exitCode=0 Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.081358 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7733a48b-2bc4-4372-a222-37bb8ea04b6d","Type":"ContainerDied","Data":"7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.081453 4778 scope.go:117] "RemoveContainer" containerID="7873b03bdc080777c3f95848a3cb2368217a2ebb6bed5cf0ae4dec3d3c66d731" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.081647 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.120385 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xht95\" (UniqueName: \"kubernetes.io/projected/7733a48b-2bc4-4372-a222-37bb8ea04b6d-kube-api-access-xht95\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.128667 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1" event={"ID":"f0341d80-4327-4c9e-bc11-0cddbc6eab66","Type":"ContainerStarted","Data":"f01be3f5a015795b5f40dbe57ae0105ae7b657a51492498554f4913e2d946654"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.128709 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1" event={"ID":"f0341d80-4327-4c9e-bc11-0cddbc6eab66","Type":"ContainerStarted","Data":"23b212b868d43a2b20b4b326a5d6e115ae1dce56fb344a9b0a15889b554a0452"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.143370 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7733a48b-2bc4-4372-a222-37bb8ea04b6d" (UID: "7733a48b-2bc4-4372-a222-37bb8ea04b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.167961 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c4d5fc-zx97x" event={"ID":"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1","Type":"ContainerStarted","Data":"580e7c05bfc29763be4beca9483310913855b264cda84343e89b99ee6458bee4"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.178606 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-config-data" (OuterVolumeSpecName: "config-data") pod "7733a48b-2bc4-4372-a222-37bb8ea04b6d" (UID: "7733a48b-2bc4-4372-a222-37bb8ea04b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.186527 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" event={"ID":"a677b5ba-f5d3-4310-ab6d-af0505e82a00","Type":"ContainerStarted","Data":"048876f254d8481a39bc4ba587f25ae5e4007ace7976831d743f8095461c0872"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.191839 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-frvxx" event={"ID":"2e1d8894-7234-40d0-b42a-9d7ab1ce638a","Type":"ContainerStarted","Data":"eb476b810cfe28d9d73622ddd41bf8c8fc415530e6ad67a1faffa32c9bd043ba"} Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.191914 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.233131 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-769c65dfd5-frvxx" podStartSLOduration=4.233116418 podStartE2EDuration="4.233116418s" podCreationTimestamp="2026-03-12 13:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:28.2328198 +0000 UTC m=+1606.681515196" watchObservedRunningTime="2026-03-12 13:36:28.233116418 +0000 UTC m=+1606.681811814" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.233387 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.241680 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7733a48b-2bc4-4372-a222-37bb8ea04b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.535462 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.562472 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.562587 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.562683 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.563868 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.563956 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" gracePeriod=600 Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.588555 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.606452 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.635261 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.635978 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-log" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.635992 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-log" Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.636007 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" containerName="nova-cell0-conductor-conductor" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.636014 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" containerName="nova-cell0-conductor-conductor" Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.636040 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-api" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.636047 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-api" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.636229 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-api" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.636253 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" containerName="nova-api-log" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.636384 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" containerName="nova-cell0-conductor-conductor" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.637058 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.641095 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.651038 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.653770 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-config-data\") pod \"37d70066-6a42-4486-a487-e27b3ab3a61b\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.653864 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d70066-6a42-4486-a487-e27b3ab3a61b-logs\") pod \"37d70066-6a42-4486-a487-e27b3ab3a61b\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.653905 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rctsf\" (UniqueName: \"kubernetes.io/projected/37d70066-6a42-4486-a487-e27b3ab3a61b-kube-api-access-rctsf\") pod \"37d70066-6a42-4486-a487-e27b3ab3a61b\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.653950 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-internal-tls-certs\") pod \"37d70066-6a42-4486-a487-e27b3ab3a61b\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.654034 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-combined-ca-bundle\") pod \"37d70066-6a42-4486-a487-e27b3ab3a61b\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.654127 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-public-tls-certs\") pod \"37d70066-6a42-4486-a487-e27b3ab3a61b\" (UID: \"37d70066-6a42-4486-a487-e27b3ab3a61b\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.672084 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37d70066-6a42-4486-a487-e27b3ab3a61b-logs" (OuterVolumeSpecName: "logs") pod "37d70066-6a42-4486-a487-e27b3ab3a61b" (UID: "37d70066-6a42-4486-a487-e27b3ab3a61b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.682151 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37d70066-6a42-4486-a487-e27b3ab3a61b-kube-api-access-rctsf" (OuterVolumeSpecName: "kube-api-access-rctsf") pod "37d70066-6a42-4486-a487-e27b3ab3a61b" (UID: "37d70066-6a42-4486-a487-e27b3ab3a61b"). InnerVolumeSpecName "kube-api-access-rctsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.707830 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.719913 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549 is running failed: container process not found" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.720329 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549 is running failed: container process not found" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.723785 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549 is running failed: container process not found" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 13:36:28 crc kubenswrapper[4778]: E0312 13:36:28.723817 4778 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9c07bd9a-becb-4422-a881-5de27a8e8e56" containerName="nova-scheduler-scheduler" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.763438 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/929bb450-949d-4f4f-9c21-de6c3fe32927-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.763594 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs2dn\" (UniqueName: \"kubernetes.io/projected/929bb450-949d-4f4f-9c21-de6c3fe32927-kube-api-access-zs2dn\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.763665 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/929bb450-949d-4f4f-9c21-de6c3fe32927-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.763721 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d70066-6a42-4486-a487-e27b3ab3a61b-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.763732 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rctsf\" (UniqueName: \"kubernetes.io/projected/37d70066-6a42-4486-a487-e27b3ab3a61b-kube-api-access-rctsf\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.793735 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.865866 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-config-data\") pod \"9c07bd9a-becb-4422-a881-5de27a8e8e56\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.866027 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-combined-ca-bundle\") pod \"9c07bd9a-becb-4422-a881-5de27a8e8e56\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.866173 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxrx6\" (UniqueName: \"kubernetes.io/projected/9c07bd9a-becb-4422-a881-5de27a8e8e56-kube-api-access-nxrx6\") pod \"9c07bd9a-becb-4422-a881-5de27a8e8e56\" (UID: \"9c07bd9a-becb-4422-a881-5de27a8e8e56\") " Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.866833 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/929bb450-949d-4f4f-9c21-de6c3fe32927-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.866997 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs2dn\" (UniqueName: \"kubernetes.io/projected/929bb450-949d-4f4f-9c21-de6c3fe32927-kube-api-access-zs2dn\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.867078 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/929bb450-949d-4f4f-9c21-de6c3fe32927-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.875926 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/929bb450-949d-4f4f-9c21-de6c3fe32927-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.900062 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/929bb450-949d-4f4f-9c21-de6c3fe32927-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.901682 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs2dn\" (UniqueName: \"kubernetes.io/projected/929bb450-949d-4f4f-9c21-de6c3fe32927-kube-api-access-zs2dn\") pod \"nova-cell0-conductor-0\" (UID: \"929bb450-949d-4f4f-9c21-de6c3fe32927\") " pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.903618 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c07bd9a-becb-4422-a881-5de27a8e8e56-kube-api-access-nxrx6" (OuterVolumeSpecName: "kube-api-access-nxrx6") pod "9c07bd9a-becb-4422-a881-5de27a8e8e56" (UID: "9c07bd9a-becb-4422-a881-5de27a8e8e56"). InnerVolumeSpecName "kube-api-access-nxrx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.910014 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-config-data" (OuterVolumeSpecName: "config-data") pod "37d70066-6a42-4486-a487-e27b3ab3a61b" (UID: "37d70066-6a42-4486-a487-e27b3ab3a61b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.910131 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37d70066-6a42-4486-a487-e27b3ab3a61b" (UID: "37d70066-6a42-4486-a487-e27b3ab3a61b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.917816 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "37d70066-6a42-4486-a487-e27b3ab3a61b" (UID: "37d70066-6a42-4486-a487-e27b3ab3a61b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.918937 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-config-data" (OuterVolumeSpecName: "config-data") pod "9c07bd9a-becb-4422-a881-5de27a8e8e56" (UID: "9c07bd9a-becb-4422-a881-5de27a8e8e56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.920832 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "37d70066-6a42-4486-a487-e27b3ab3a61b" (UID: "37d70066-6a42-4486-a487-e27b3ab3a61b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.946583 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c07bd9a-becb-4422-a881-5de27a8e8e56" (UID: "9c07bd9a-becb-4422-a881-5de27a8e8e56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970663 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970712 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970727 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970738 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970749 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c07bd9a-becb-4422-a881-5de27a8e8e56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970760 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxrx6\" (UniqueName: \"kubernetes.io/projected/9c07bd9a-becb-4422-a881-5de27a8e8e56-kube-api-access-nxrx6\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.970774 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37d70066-6a42-4486-a487-e27b3ab3a61b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:28 crc kubenswrapper[4778]: I0312 13:36:28.982976 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.204570 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37d70066-6a42-4486-a487-e27b3ab3a61b","Type":"ContainerDied","Data":"c412c8241a59192093777ad55c60d09316d57b1f207c8116b8342fac0e609d85"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.204913 4778 scope.go:117] "RemoveContainer" containerID="6fb0ea63ecde6cfac6694eb778a0f0043874e52ed36561d1c373be870defe193" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.204836 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.218282 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"5a29f7b0-d851-4967-802b-91e301ce82f2","Type":"ContainerStarted","Data":"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.218353 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"5a29f7b0-d851-4967-802b-91e301ce82f2","Type":"ContainerStarted","Data":"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.222471 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1" event={"ID":"f0341d80-4327-4c9e-bc11-0cddbc6eab66","Type":"ContainerStarted","Data":"8c2261933274af98452e3406c6a1491fa7547083d2a42fd8ebde0430b36e07c4"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.225494 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c4d5fc-zx97x" event={"ID":"8a67d4b7-d8eb-40f4-b51d-62e92c6042c1","Type":"ContainerStarted","Data":"0a03291c72a2f101c8185bc5efeea52f3870fb1b63ef8d6feb2ae00287399781"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.226053 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.228266 4778 generic.go:334] "Generic (PLEG): container finished" podID="9c07bd9a-becb-4422-a881-5de27a8e8e56" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" exitCode=0 Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.228337 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c07bd9a-becb-4422-a881-5de27a8e8e56","Type":"ContainerDied","Data":"76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.228361 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c07bd9a-becb-4422-a881-5de27a8e8e56","Type":"ContainerDied","Data":"7c66f4ab481a2ffe8aa1f978637b76237ffd5a0742d58abdd29d2a665dd400d8"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.228429 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.235084 4778 generic.go:334] "Generic (PLEG): container finished" podID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerID="0a63a4d47752d25e9f6c0d6aa9ed71121a4afe876250e6e10c1c1091bf2b8d8f" exitCode=0 Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.235352 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" event={"ID":"a677b5ba-f5d3-4310-ab6d-af0505e82a00","Type":"ContainerDied","Data":"0a63a4d47752d25e9f6c0d6aa9ed71121a4afe876250e6e10c1c1091bf2b8d8f"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.253392 4778 scope.go:117] "RemoveContainer" containerID="2b4aa266ad205b7e0a6d8899547a75fe40c64017eb43d076bced61bb7cc36c19" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.256285 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-frvxx" event={"ID":"2e1d8894-7234-40d0-b42a-9d7ab1ce638a","Type":"ContainerStarted","Data":"1e1f7bdc1b1f277f7f73a02d00233e5689179238b97d5a12c4ec486b0a81ef94"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.262921 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-metadata-0" podStartSLOduration=4.262895318 podStartE2EDuration="4.262895318s" podCreationTimestamp="2026-03-12 13:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:29.244433244 +0000 UTC m=+1607.693128660" watchObservedRunningTime="2026-03-12 13:36:29.262895318 +0000 UTC m=+1607.711590734" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.298730 4778 generic.go:334] "Generic (PLEG): container finished" podID="7be5cd74-51aa-4be4-bee0-bcd4414e988c" containerID="f7abc5ad094a49286dcbc9e9529cdeaa0cf757440a3faf4823fd72045e913f36" exitCode=0 Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.299278 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6796c46585-tk69s" event={"ID":"7be5cd74-51aa-4be4-bee0-bcd4414e988c","Type":"ContainerDied","Data":"f7abc5ad094a49286dcbc9e9529cdeaa0cf757440a3faf4823fd72045e913f36"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.336866 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.338458 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" exitCode=0 Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.338564 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e"} Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.339177 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:36:29 crc kubenswrapper[4778]: E0312 13:36:29.339504 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.355764 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.402154 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-1" podStartSLOduration=4.402136247 podStartE2EDuration="4.402136247s" podCreationTimestamp="2026-03-12 13:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:29.322266142 +0000 UTC m=+1607.770961558" watchObservedRunningTime="2026-03-12 13:36:29.402136247 +0000 UTC m=+1607.850831643" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.402277 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: E0312 13:36:29.402905 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c07bd9a-becb-4422-a881-5de27a8e8e56" containerName="nova-scheduler-scheduler" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.402942 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c07bd9a-becb-4422-a881-5de27a8e8e56" containerName="nova-scheduler-scheduler" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.403251 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c07bd9a-becb-4422-a881-5de27a8e8e56" containerName="nova-scheduler-scheduler" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.404640 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.492956 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-566c4d5fc-zx97x" podStartSLOduration=5.492934343 podStartE2EDuration="5.492934343s" podCreationTimestamp="2026-03-12 13:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:29.353900979 +0000 UTC m=+1607.802596375" watchObservedRunningTime="2026-03-12 13:36:29.492934343 +0000 UTC m=+1607.941629729" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.495318 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.495960 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.496155 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np6gx\" (UniqueName: \"kubernetes.io/projected/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-kube-api-access-np6gx\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.497299 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-logs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.497483 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-config-data\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.498056 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-public-tls-certs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.504028 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.587453 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.601058 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-public-tls-certs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.601135 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.601173 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.601215 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np6gx\" (UniqueName: \"kubernetes.io/projected/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-kube-api-access-np6gx\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.601324 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-logs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.601371 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-config-data\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.608879 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-config-data\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.610484 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.615101 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.616135 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-logs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.620771 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.624036 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-public-tls-certs\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.624828 4778 scope.go:117] "RemoveContainer" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.640216 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np6gx\" (UniqueName: \"kubernetes.io/projected/13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4-kube-api-access-np6gx\") pod \"nova-api-0\" (UID: \"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4\") " pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.640721 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.643056 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.646753 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.655509 4778 scope.go:117] "RemoveContainer" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" Mar 12 13:36:29 crc kubenswrapper[4778]: E0312 13:36:29.657110 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549\": container with ID starting with 76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549 not found: ID does not exist" containerID="76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.657147 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549"} err="failed to get container status \"76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549\": rpc error: code = NotFound desc = could not find container \"76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549\": container with ID starting with 76612676e8ec4eca2c187a9ab03eca1247f93e79ade8e00d7568f9fdf3aca549 not found: ID does not exist" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.657167 4778 scope.go:117] "RemoveContainer" containerID="572aad6c3b1a3f7c9ef45b8b4feb0d367e7e7916d0ab8dd064e2b8ac87268c51" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.703975 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f613745b-fe33-4918-9e0a-da2a59c55e33-config-data\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.704025 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f613745b-fe33-4918-9e0a-da2a59c55e33-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.704046 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnx9t\" (UniqueName: \"kubernetes.io/projected/f613745b-fe33-4918-9e0a-da2a59c55e33-kube-api-access-fnx9t\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.713127 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.767779 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.786094 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.806391 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f613745b-fe33-4918-9e0a-da2a59c55e33-config-data\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.806449 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f613745b-fe33-4918-9e0a-da2a59c55e33-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.806772 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnx9t\" (UniqueName: \"kubernetes.io/projected/f613745b-fe33-4918-9e0a-da2a59c55e33-kube-api-access-fnx9t\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.825758 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f613745b-fe33-4918-9e0a-da2a59c55e33-config-data\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.827147 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f613745b-fe33-4918-9e0a-da2a59c55e33-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.835302 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnx9t\" (UniqueName: \"kubernetes.io/projected/f613745b-fe33-4918-9e0a-da2a59c55e33-kube-api-access-fnx9t\") pod \"nova-scheduler-0\" (UID: \"f613745b-fe33-4918-9e0a-da2a59c55e33\") " pod="openstack/nova-scheduler-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.887301 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.953894 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:29 crc kubenswrapper[4778]: I0312 13:36:29.970206 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.025280 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-config\") pod \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.025376 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc9pp\" (UniqueName: \"kubernetes.io/projected/7be5cd74-51aa-4be4-bee0-bcd4414e988c-kube-api-access-mc9pp\") pod \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.025624 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-swift-storage-0\") pod \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.025807 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-sb\") pod \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.025912 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-nb\") pod \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.026071 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-svc\") pod \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\" (UID: \"7be5cd74-51aa-4be4-bee0-bcd4414e988c\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.045908 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be5cd74-51aa-4be4-bee0-bcd4414e988c-kube-api-access-mc9pp" (OuterVolumeSpecName: "kube-api-access-mc9pp") pod "7be5cd74-51aa-4be4-bee0-bcd4414e988c" (UID: "7be5cd74-51aa-4be4-bee0-bcd4414e988c"). InnerVolumeSpecName "kube-api-access-mc9pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.134596 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc9pp\" (UniqueName: \"kubernetes.io/projected/7be5cd74-51aa-4be4-bee0-bcd4414e988c-kube-api-access-mc9pp\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.232085 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-config" (OuterVolumeSpecName: "config") pod "7be5cd74-51aa-4be4-bee0-bcd4414e988c" (UID: "7be5cd74-51aa-4be4-bee0-bcd4414e988c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.237139 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.249351 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7be5cd74-51aa-4be4-bee0-bcd4414e988c" (UID: "7be5cd74-51aa-4be4-bee0-bcd4414e988c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.261053 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7be5cd74-51aa-4be4-bee0-bcd4414e988c" (UID: "7be5cd74-51aa-4be4-bee0-bcd4414e988c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.261092 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7be5cd74-51aa-4be4-bee0-bcd4414e988c" (UID: "7be5cd74-51aa-4be4-bee0-bcd4414e988c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.267157 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7be5cd74-51aa-4be4-bee0-bcd4414e988c" (UID: "7be5cd74-51aa-4be4-bee0-bcd4414e988c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.296111 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37d70066-6a42-4486-a487-e27b3ab3a61b" path="/var/lib/kubelet/pods/37d70066-6a42-4486-a487-e27b3ab3a61b/volumes" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.297648 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7733a48b-2bc4-4372-a222-37bb8ea04b6d" path="/var/lib/kubelet/pods/7733a48b-2bc4-4372-a222-37bb8ea04b6d/volumes" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.298371 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c07bd9a-becb-4422-a881-5de27a8e8e56" path="/var/lib/kubelet/pods/9c07bd9a-becb-4422-a881-5de27a8e8e56/volumes" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.344209 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.344248 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.344262 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.344273 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7be5cd74-51aa-4be4-bee0-bcd4414e988c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.365585 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.375699 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" event={"ID":"a677b5ba-f5d3-4310-ab6d-af0505e82a00","Type":"ContainerStarted","Data":"80b9a94e51ace133a39bb4f360454c37e2be50602309d428d0792de3b24d2efc"} Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.377374 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.380017 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"929bb450-949d-4f4f-9c21-de6c3fe32927","Type":"ContainerStarted","Data":"52042b0769bcccf846e369aa04436e8cee2985c96f965a205aee13cc5d76845b"} Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.380051 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"929bb450-949d-4f4f-9c21-de6c3fe32927","Type":"ContainerStarted","Data":"a745c0714a3158f1d8c67ba0596d6c8fcebec865205de5a0ab80dcef2a2db29a"} Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.380990 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.385377 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6796c46585-tk69s" event={"ID":"7be5cd74-51aa-4be4-bee0-bcd4414e988c","Type":"ContainerDied","Data":"9611fdd9f50898df9bc39508d9d61ff6e00eab70329522bacb24a65cad5f58f2"} Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.385430 4778 scope.go:117] "RemoveContainer" containerID="f7abc5ad094a49286dcbc9e9529cdeaa0cf757440a3faf4823fd72045e913f36" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.385517 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6796c46585-tk69s" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.401469 4778 generic.go:334] "Generic (PLEG): container finished" podID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerID="1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed" exitCode=0 Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.401831 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5f2bac2-0571-44d8-ba4e-c006600506a5","Type":"ContainerDied","Data":"1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed"} Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.401883 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5f2bac2-0571-44d8-ba4e-c006600506a5","Type":"ContainerDied","Data":"63ddaf1ae0c88152b73f19a7bbf611a71857e4349ca72eb4c73d9d4e815e1b3c"} Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.401976 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.416233 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" podStartSLOduration=4.416214803 podStartE2EDuration="4.416214803s" podCreationTimestamp="2026-03-12 13:36:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:30.407606679 +0000 UTC m=+1608.856302075" watchObservedRunningTime="2026-03-12 13:36:30.416214803 +0000 UTC m=+1608.864910199" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.446814 4778 scope.go:117] "RemoveContainer" containerID="1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.447969 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-config-data\") pod \"e5f2bac2-0571-44d8-ba4e-c006600506a5\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.448016 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5dsf\" (UniqueName: \"kubernetes.io/projected/e5f2bac2-0571-44d8-ba4e-c006600506a5-kube-api-access-d5dsf\") pod \"e5f2bac2-0571-44d8-ba4e-c006600506a5\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.448175 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2bac2-0571-44d8-ba4e-c006600506a5-logs\") pod \"e5f2bac2-0571-44d8-ba4e-c006600506a5\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.448221 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-combined-ca-bundle\") pod \"e5f2bac2-0571-44d8-ba4e-c006600506a5\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.448307 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-nova-metadata-tls-certs\") pod \"e5f2bac2-0571-44d8-ba4e-c006600506a5\" (UID: \"e5f2bac2-0571-44d8-ba4e-c006600506a5\") " Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.448927 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5f2bac2-0571-44d8-ba4e-c006600506a5-logs" (OuterVolumeSpecName: "logs") pod "e5f2bac2-0571-44d8-ba4e-c006600506a5" (UID: "e5f2bac2-0571-44d8-ba4e-c006600506a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.462018 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f2bac2-0571-44d8-ba4e-c006600506a5-kube-api-access-d5dsf" (OuterVolumeSpecName: "kube-api-access-d5dsf") pod "e5f2bac2-0571-44d8-ba4e-c006600506a5" (UID: "e5f2bac2-0571-44d8-ba4e-c006600506a5"). InnerVolumeSpecName "kube-api-access-d5dsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.483251 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.483223214 podStartE2EDuration="2.483223214s" podCreationTimestamp="2026-03-12 13:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:30.449404734 +0000 UTC m=+1608.898100130" watchObservedRunningTime="2026-03-12 13:36:30.483223214 +0000 UTC m=+1608.931918620" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.516605 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6796c46585-tk69s"] Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.518899 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-config-data" (OuterVolumeSpecName: "config-data") pod "e5f2bac2-0571-44d8-ba4e-c006600506a5" (UID: "e5f2bac2-0571-44d8-ba4e-c006600506a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.527311 4778 scope.go:117] "RemoveContainer" containerID="3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.533845 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6796c46585-tk69s"] Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.538441 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5f2bac2-0571-44d8-ba4e-c006600506a5" (UID: "e5f2bac2-0571-44d8-ba4e-c006600506a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.553731 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2bac2-0571-44d8-ba4e-c006600506a5-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.553762 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.553773 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.553783 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5dsf\" (UniqueName: \"kubernetes.io/projected/e5f2bac2-0571-44d8-ba4e-c006600506a5-kube-api-access-d5dsf\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.568352 4778 scope.go:117] "RemoveContainer" containerID="1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.568629 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e5f2bac2-0571-44d8-ba4e-c006600506a5" (UID: "e5f2bac2-0571-44d8-ba4e-c006600506a5"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:30 crc kubenswrapper[4778]: E0312 13:36:30.569744 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed\": container with ID starting with 1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed not found: ID does not exist" containerID="1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.569793 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed"} err="failed to get container status \"1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed\": rpc error: code = NotFound desc = could not find container \"1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed\": container with ID starting with 1938dab200355bd40396968513a110679aaccecc8babb8b2a8c4c460989f58ed not found: ID does not exist" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.569824 4778 scope.go:117] "RemoveContainer" containerID="3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c" Mar 12 13:36:30 crc kubenswrapper[4778]: E0312 13:36:30.570643 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c\": container with ID starting with 3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c not found: ID does not exist" containerID="3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.570682 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c"} err="failed to get container status \"3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c\": rpc error: code = NotFound desc = could not find container \"3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c\": container with ID starting with 3f33e9e86bb02fc44b869bbdad27f6457624a82e4496ad8d8db76de0c3d1fb4c not found: ID does not exist" Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.620135 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 13:36:30 crc kubenswrapper[4778]: I0312 13:36:30.657604 4778 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f2bac2-0571-44d8-ba4e-c006600506a5-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:31 crc kubenswrapper[4778]: E0312 13:36:31.114790 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 12 13:36:31 crc kubenswrapper[4778]: E0312 13:36:31.128323 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 12 13:36:31 crc kubenswrapper[4778]: E0312 13:36:31.132343 4778 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 12 13:36:31 crc kubenswrapper[4778]: E0312 13:36:31.132419 4778 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" containerName="nova-cell1-conductor-conductor" Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.184114 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 13:36:31 crc kubenswrapper[4778]: W0312 13:36:31.186321 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13b8e1df_5a8c_44de_b8e8_6c7efdb8bad4.slice/crio-be7e1bcb037e22cbdbae44d248064a96e8822fc40563d6af814e4980aca9133c WatchSource:0}: Error finding container be7e1bcb037e22cbdbae44d248064a96e8822fc40563d6af814e4980aca9133c: Status 404 returned error can't find the container with id be7e1bcb037e22cbdbae44d248064a96e8822fc40563d6af814e4980aca9133c Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.364155 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.374668 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.442499 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f613745b-fe33-4918-9e0a-da2a59c55e33","Type":"ContainerStarted","Data":"8a09c7c89f2aff34ffefc779797752345605e724b52f5e03391e22b5822ed81a"} Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.443115 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f613745b-fe33-4918-9e0a-da2a59c55e33","Type":"ContainerStarted","Data":"f9fc1f2571ef8dc3431c1ecf1910fa61e4e09394d37fc13345a23eda1f9922e1"} Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.456645 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4","Type":"ContainerStarted","Data":"aa09b24193825e4751e820dec9702bffb5238df9786e2014a8c5502a2a867ae7"} Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.456719 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4","Type":"ContainerStarted","Data":"be7e1bcb037e22cbdbae44d248064a96e8822fc40563d6af814e4980aca9133c"} Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.465137 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-metadata-0" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-log" containerID="cri-o://417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0" gracePeriod=30 Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.467644 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-metadata-0" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-metadata" containerID="cri-o://dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961" gracePeriod=30 Mar 12 13:36:31 crc kubenswrapper[4778]: I0312 13:36:31.551904 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.551871677 podStartE2EDuration="2.551871677s" podCreationTimestamp="2026-03-12 13:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:31.461341479 +0000 UTC m=+1609.910036885" watchObservedRunningTime="2026-03-12 13:36:31.551871677 +0000 UTC m=+1610.000567073" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.042919 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.149016 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4cqk\" (UniqueName: \"kubernetes.io/projected/5a29f7b0-d851-4967-802b-91e301ce82f2-kube-api-access-m4cqk\") pod \"5a29f7b0-d851-4967-802b-91e301ce82f2\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.149356 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a29f7b0-d851-4967-802b-91e301ce82f2-logs\") pod \"5a29f7b0-d851-4967-802b-91e301ce82f2\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.149504 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-combined-ca-bundle\") pod \"5a29f7b0-d851-4967-802b-91e301ce82f2\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.149616 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-config-data\") pod \"5a29f7b0-d851-4967-802b-91e301ce82f2\" (UID: \"5a29f7b0-d851-4967-802b-91e301ce82f2\") " Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.149983 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a29f7b0-d851-4967-802b-91e301ce82f2-logs" (OuterVolumeSpecName: "logs") pod "5a29f7b0-d851-4967-802b-91e301ce82f2" (UID: "5a29f7b0-d851-4967-802b-91e301ce82f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.150380 4778 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a29f7b0-d851-4967-802b-91e301ce82f2-logs\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.157419 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a29f7b0-d851-4967-802b-91e301ce82f2-kube-api-access-m4cqk" (OuterVolumeSpecName: "kube-api-access-m4cqk") pod "5a29f7b0-d851-4967-802b-91e301ce82f2" (UID: "5a29f7b0-d851-4967-802b-91e301ce82f2"). InnerVolumeSpecName "kube-api-access-m4cqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.183385 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a29f7b0-d851-4967-802b-91e301ce82f2" (UID: "5a29f7b0-d851-4967-802b-91e301ce82f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.187400 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-config-data" (OuterVolumeSpecName: "config-data") pod "5a29f7b0-d851-4967-802b-91e301ce82f2" (UID: "5a29f7b0-d851-4967-802b-91e301ce82f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.253455 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.253881 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a29f7b0-d851-4967-802b-91e301ce82f2-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.253898 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4cqk\" (UniqueName: \"kubernetes.io/projected/5a29f7b0-d851-4967-802b-91e301ce82f2-kube-api-access-m4cqk\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.277175 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7be5cd74-51aa-4be4-bee0-bcd4414e988c" path="/var/lib/kubelet/pods/7be5cd74-51aa-4be4-bee0-bcd4414e988c/volumes" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.278277 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" path="/var/lib/kubelet/pods/e5f2bac2-0571-44d8-ba4e-c006600506a5/volumes" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.480902 4778 generic.go:334] "Generic (PLEG): container finished" podID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerID="dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961" exitCode=0 Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.481266 4778 generic.go:334] "Generic (PLEG): container finished" podID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerID="417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0" exitCode=143 Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.481128 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"5a29f7b0-d851-4967-802b-91e301ce82f2","Type":"ContainerDied","Data":"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961"} Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.481234 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.481376 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"5a29f7b0-d851-4967-802b-91e301ce82f2","Type":"ContainerDied","Data":"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0"} Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.481405 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"5a29f7b0-d851-4967-802b-91e301ce82f2","Type":"ContainerDied","Data":"7686b13a63fb7303c82944b69b0f74a27bb498a86b8a8900db3aa379dab6a697"} Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.481426 4778 scope.go:117] "RemoveContainer" containerID="dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.484584 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4","Type":"ContainerStarted","Data":"a29462829c4d1d056fbfecca3676d344fa30407641f877d1b84a8a5c398c1120"} Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.509722 4778 scope.go:117] "RemoveContainer" containerID="417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.524385 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.524358582 podStartE2EDuration="3.524358582s" podCreationTimestamp="2026-03-12 13:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:32.514655236 +0000 UTC m=+1610.963350632" watchObservedRunningTime="2026-03-12 13:36:32.524358582 +0000 UTC m=+1610.973053988" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.544987 4778 scope.go:117] "RemoveContainer" containerID="dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961" Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.545545 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961\": container with ID starting with dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961 not found: ID does not exist" containerID="dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.545589 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961"} err="failed to get container status \"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961\": rpc error: code = NotFound desc = could not find container \"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961\": container with ID starting with dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961 not found: ID does not exist" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.545621 4778 scope.go:117] "RemoveContainer" containerID="417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0" Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.545864 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0\": container with ID starting with 417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0 not found: ID does not exist" containerID="417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.545894 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0"} err="failed to get container status \"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0\": rpc error: code = NotFound desc = could not find container \"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0\": container with ID starting with 417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0 not found: ID does not exist" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.545918 4778 scope.go:117] "RemoveContainer" containerID="dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.546151 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961"} err="failed to get container status \"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961\": rpc error: code = NotFound desc = could not find container \"dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961\": container with ID starting with dd6bb746e6fc0c601d6ae2bcf58c264b68fb62b2faf56722bc914a19843b5961 not found: ID does not exist" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.546191 4778 scope.go:117] "RemoveContainer" containerID="417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.546443 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0"} err="failed to get container status \"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0\": rpc error: code = NotFound desc = could not find container \"417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0\": container with ID starting with 417f17c877a12302e92668f628e55146a1d2400c3af2e99761c6921a618abcb0 not found: ID does not exist" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.555259 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.576442 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.594700 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.595388 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-log" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595410 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-log" Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.595435 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-metadata" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595441 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-metadata" Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.595466 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-metadata" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595476 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-metadata" Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.595489 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be5cd74-51aa-4be4-bee0-bcd4414e988c" containerName="init" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595495 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be5cd74-51aa-4be4-bee0-bcd4414e988c" containerName="init" Mar 12 13:36:32 crc kubenswrapper[4778]: E0312 13:36:32.595515 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-log" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595521 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-log" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595739 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-metadata" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595758 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-metadata" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595765 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f2bac2-0571-44d8-ba4e-c006600506a5" containerName="nova-metadata-log" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595775 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be5cd74-51aa-4be4-bee0-bcd4414e988c" containerName="init" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.595789 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" containerName="nova-cell1-metadata-log" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.597243 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.601885 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-cell1-internal-svc" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.602176 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-metadata-config-data" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.606477 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.668834 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-config-data\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.668893 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z95k8\" (UniqueName: \"kubernetes.io/projected/c289a520-78eb-433f-b7a4-0c03be917c18-kube-api-access-z95k8\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.668963 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-nova-metadata-tls-certs\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.669014 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c289a520-78eb-433f-b7a4-0c03be917c18-logs\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.669073 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-combined-ca-bundle\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.772258 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-config-data\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.772325 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z95k8\" (UniqueName: \"kubernetes.io/projected/c289a520-78eb-433f-b7a4-0c03be917c18-kube-api-access-z95k8\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.772375 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-nova-metadata-tls-certs\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.772430 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c289a520-78eb-433f-b7a4-0c03be917c18-logs\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.772469 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-combined-ca-bundle\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.773605 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c289a520-78eb-433f-b7a4-0c03be917c18-logs\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.780825 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-nova-metadata-tls-certs\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.780857 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-combined-ca-bundle\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.781149 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c289a520-78eb-433f-b7a4-0c03be917c18-config-data\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.796715 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z95k8\" (UniqueName: \"kubernetes.io/projected/c289a520-78eb-433f-b7a4-0c03be917c18-kube-api-access-z95k8\") pod \"nova-cell1-metadata-0\" (UID: \"c289a520-78eb-433f-b7a4-0c03be917c18\") " pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:32 crc kubenswrapper[4778]: I0312 13:36:32.916295 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:33 crc kubenswrapper[4778]: I0312 13:36:33.496560 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-metadata-0"] Mar 12 13:36:33 crc kubenswrapper[4778]: I0312 13:36:33.504021 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"c289a520-78eb-433f-b7a4-0c03be917c18","Type":"ContainerStarted","Data":"f2aa1bb4f48d2cacad481df82ac182e97844b78757aa2641bc74a3d67fa66465"} Mar 12 13:36:34 crc kubenswrapper[4778]: I0312 13:36:34.296354 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a29f7b0-d851-4967-802b-91e301ce82f2" path="/var/lib/kubelet/pods/5a29f7b0-d851-4967-802b-91e301ce82f2/volumes" Mar 12 13:36:34 crc kubenswrapper[4778]: I0312 13:36:34.523564 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"c289a520-78eb-433f-b7a4-0c03be917c18","Type":"ContainerStarted","Data":"2cf98c7e55bb9691ad7fd575db56e4f9545613bf3af1b6fb6fc96b703ff2b197"} Mar 12 13:36:34 crc kubenswrapper[4778]: I0312 13:36:34.523618 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-metadata-0" event={"ID":"c289a520-78eb-433f-b7a4-0c03be917c18","Type":"ContainerStarted","Data":"9b59dda90b6bbb6f76cb19e3620376e472a849789bd2baf7523bd43526b2a527"} Mar 12 13:36:34 crc kubenswrapper[4778]: I0312 13:36:34.971260 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 13:36:35 crc kubenswrapper[4778]: I0312 13:36:35.541888 4778 generic.go:334] "Generic (PLEG): container finished" podID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" containerID="17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f" exitCode=0 Mar 12 13:36:35 crc kubenswrapper[4778]: I0312 13:36:35.543086 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e28e8bc2-4b60-447e-b78e-99f53f0559e9","Type":"ContainerDied","Data":"17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f"} Mar 12 13:36:35 crc kubenswrapper[4778]: I0312 13:36:35.999398 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.021618 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-metadata-0" podStartSLOduration=4.021595754 podStartE2EDuration="4.021595754s" podCreationTimestamp="2026-03-12 13:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:34.551683009 +0000 UTC m=+1613.000378405" watchObservedRunningTime="2026-03-12 13:36:36.021595754 +0000 UTC m=+1614.470291140" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.052062 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srnlz\" (UniqueName: \"kubernetes.io/projected/e28e8bc2-4b60-447e-b78e-99f53f0559e9-kube-api-access-srnlz\") pod \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.052139 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-combined-ca-bundle\") pod \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.053239 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-config-data\") pod \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\" (UID: \"e28e8bc2-4b60-447e-b78e-99f53f0559e9\") " Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.078487 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e28e8bc2-4b60-447e-b78e-99f53f0559e9-kube-api-access-srnlz" (OuterVolumeSpecName: "kube-api-access-srnlz") pod "e28e8bc2-4b60-447e-b78e-99f53f0559e9" (UID: "e28e8bc2-4b60-447e-b78e-99f53f0559e9"). InnerVolumeSpecName "kube-api-access-srnlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.097526 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e28e8bc2-4b60-447e-b78e-99f53f0559e9" (UID: "e28e8bc2-4b60-447e-b78e-99f53f0559e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.104565 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-config-data" (OuterVolumeSpecName: "config-data") pod "e28e8bc2-4b60-447e-b78e-99f53f0559e9" (UID: "e28e8bc2-4b60-447e-b78e-99f53f0559e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.156312 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srnlz\" (UniqueName: \"kubernetes.io/projected/e28e8bc2-4b60-447e-b78e-99f53f0559e9-kube-api-access-srnlz\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.156354 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.156365 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28e8bc2-4b60-447e-b78e-99f53f0559e9-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.248500 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-1" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.248558 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-1" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.556454 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e28e8bc2-4b60-447e-b78e-99f53f0559e9","Type":"ContainerDied","Data":"0b4aff5eb3ee6cc75fbeaaa57c05dff4153b4e03f714593a99c2f4d9aa7da572"} Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.556963 4778 scope.go:117] "RemoveContainer" containerID="17f6ecc58bfeead13bd408fa3389fcd5b9ea0127020d364f507d2277de0d4c6f" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.556783 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.592301 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.608150 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.632456 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:36:36 crc kubenswrapper[4778]: E0312 13:36:36.633250 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" containerName="nova-cell1-conductor-conductor" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.633271 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" containerName="nova-cell1-conductor-conductor" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.633585 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" containerName="nova-cell1-conductor-conductor" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.634724 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.637864 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.651600 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.666379 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1466aea3-fa10-49a6-a254-a96a52091aca-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.667221 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1466aea3-fa10-49a6-a254-a96a52091aca-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.667426 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vpwx\" (UniqueName: \"kubernetes.io/projected/1466aea3-fa10-49a6-a254-a96a52091aca-kube-api-access-9vpwx\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.767583 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.769640 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1466aea3-fa10-49a6-a254-a96a52091aca-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.769733 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vpwx\" (UniqueName: \"kubernetes.io/projected/1466aea3-fa10-49a6-a254-a96a52091aca-kube-api-access-9vpwx\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.769814 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1466aea3-fa10-49a6-a254-a96a52091aca-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.775689 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1466aea3-fa10-49a6-a254-a96a52091aca-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.783126 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1466aea3-fa10-49a6-a254-a96a52091aca-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.800660 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vpwx\" (UniqueName: \"kubernetes.io/projected/1466aea3-fa10-49a6-a254-a96a52091aca-kube-api-access-9vpwx\") pod \"nova-cell1-conductor-0\" (UID: \"1466aea3-fa10-49a6-a254-a96a52091aca\") " pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.869668 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vbzn5"] Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.869913 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="dnsmasq-dns" containerID="cri-o://9226d052c31f98b5c3da17ce19bbc81e718b949c212eab5fa79f7c540fdf830a" gracePeriod=10 Mar 12 13:36:36 crc kubenswrapper[4778]: I0312 13:36:36.960479 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:37 crc kubenswrapper[4778]: I0312 13:36:37.261357 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-1" podUID="f0341d80-4327-4c9e-bc11-0cddbc6eab66" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.216:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:36:37 crc kubenswrapper[4778]: I0312 13:36:37.262311 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-1" podUID="f0341d80-4327-4c9e-bc11-0cddbc6eab66" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.216:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:36:37 crc kubenswrapper[4778]: I0312 13:36:37.639028 4778 generic.go:334] "Generic (PLEG): container finished" podID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerID="9226d052c31f98b5c3da17ce19bbc81e718b949c212eab5fa79f7c540fdf830a" exitCode=0 Mar 12 13:36:37 crc kubenswrapper[4778]: I0312 13:36:37.639203 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" event={"ID":"d621990b-b3fb-457c-a7b8-0726fa89a5e6","Type":"ContainerDied","Data":"9226d052c31f98b5c3da17ce19bbc81e718b949c212eab5fa79f7c540fdf830a"} Mar 12 13:36:37 crc kubenswrapper[4778]: I0312 13:36:37.996619 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.108899 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.133577 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-svc\") pod \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.133744 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-config\") pod \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.133867 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-swift-storage-0\") pod \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.133898 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5ccj\" (UniqueName: \"kubernetes.io/projected/d621990b-b3fb-457c-a7b8-0726fa89a5e6-kube-api-access-b5ccj\") pod \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.134179 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-nb\") pod \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.134241 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-sb\") pod \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\" (UID: \"d621990b-b3fb-457c-a7b8-0726fa89a5e6\") " Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.176323 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d621990b-b3fb-457c-a7b8-0726fa89a5e6-kube-api-access-b5ccj" (OuterVolumeSpecName: "kube-api-access-b5ccj") pod "d621990b-b3fb-457c-a7b8-0726fa89a5e6" (UID: "d621990b-b3fb-457c-a7b8-0726fa89a5e6"). InnerVolumeSpecName "kube-api-access-b5ccj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.223148 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d621990b-b3fb-457c-a7b8-0726fa89a5e6" (UID: "d621990b-b3fb-457c-a7b8-0726fa89a5e6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.238057 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d621990b-b3fb-457c-a7b8-0726fa89a5e6" (UID: "d621990b-b3fb-457c-a7b8-0726fa89a5e6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.238650 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5ccj\" (UniqueName: \"kubernetes.io/projected/d621990b-b3fb-457c-a7b8-0726fa89a5e6-kube-api-access-b5ccj\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.238693 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.238707 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.245174 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d621990b-b3fb-457c-a7b8-0726fa89a5e6" (UID: "d621990b-b3fb-457c-a7b8-0726fa89a5e6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.267626 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-config" (OuterVolumeSpecName: "config") pod "d621990b-b3fb-457c-a7b8-0726fa89a5e6" (UID: "d621990b-b3fb-457c-a7b8-0726fa89a5e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.279467 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e28e8bc2-4b60-447e-b78e-99f53f0559e9" path="/var/lib/kubelet/pods/e28e8bc2-4b60-447e-b78e-99f53f0559e9/volumes" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.280059 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d621990b-b3fb-457c-a7b8-0726fa89a5e6" (UID: "d621990b-b3fb-457c-a7b8-0726fa89a5e6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.346480 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.346523 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.346666 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621990b-b3fb-457c-a7b8-0726fa89a5e6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.655009 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" event={"ID":"d621990b-b3fb-457c-a7b8-0726fa89a5e6","Type":"ContainerDied","Data":"7c046518ad4ee249311d20eb84f556ea55869944e1e9d121bc2b448648522cec"} Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.655070 4778 scope.go:117] "RemoveContainer" containerID="9226d052c31f98b5c3da17ce19bbc81e718b949c212eab5fa79f7c540fdf830a" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.655275 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.658397 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1466aea3-fa10-49a6-a254-a96a52091aca","Type":"ContainerStarted","Data":"70d875f5a4aa3f21adec3e4eb646a10acf69a7927b64f60ffab4c71759da535a"} Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.658473 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1466aea3-fa10-49a6-a254-a96a52091aca","Type":"ContainerStarted","Data":"1f27b5dc0babf6d3d07bf35ec2c31830dbbb5b81af3caf00127f2f0c76025e27"} Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.660002 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.693756 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vbzn5"] Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.694922 4778 scope.go:117] "RemoveContainer" containerID="f768634e6581a58404932d5b274b7e499ff8a446926b77d44c652d5c4c0bad66" Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.715767 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vbzn5"] Mar 12 13:36:38 crc kubenswrapper[4778]: I0312 13:36:38.720086 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.720060059 podStartE2EDuration="2.720060059s" podCreationTimestamp="2026-03-12 13:36:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:38.703323844 +0000 UTC m=+1617.152019230" watchObservedRunningTime="2026-03-12 13:36:38.720060059 +0000 UTC m=+1617.168755455" Mar 12 13:36:39 crc kubenswrapper[4778]: I0312 13:36:39.213884 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 12 13:36:39 crc kubenswrapper[4778]: I0312 13:36:39.888160 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:36:39 crc kubenswrapper[4778]: I0312 13:36:39.889568 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 13:36:40 crc kubenswrapper[4778]: I0312 13:36:40.273256 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 12 13:36:40 crc kubenswrapper[4778]: I0312 13:36:40.351970 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" path="/var/lib/kubelet/pods/d621990b-b3fb-457c-a7b8-0726fa89a5e6/volumes" Mar 12 13:36:40 crc kubenswrapper[4778]: I0312 13:36:40.367749 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 12 13:36:41 crc kubenswrapper[4778]: I0312 13:36:41.041525 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.221:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:36:41 crc kubenswrapper[4778]: I0312 13:36:41.041819 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.221:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:36:41 crc kubenswrapper[4778]: I0312 13:36:41.090002 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 12 13:36:42 crc kubenswrapper[4778]: I0312 13:36:42.263888 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:36:42 crc kubenswrapper[4778]: E0312 13:36:42.264860 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:36:42 crc kubenswrapper[4778]: I0312 13:36:42.694135 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c5cd4d5-vbzn5" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.204:5353: i/o timeout" Mar 12 13:36:42 crc kubenswrapper[4778]: I0312 13:36:42.917780 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:42 crc kubenswrapper[4778]: I0312 13:36:42.917831 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:43 crc kubenswrapper[4778]: I0312 13:36:43.936422 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-cell1-metadata-0" podUID="c289a520-78eb-433f-b7a4-0c03be917c18" containerName="nova-cell1-metadata-log" probeResult="failure" output="Get \"https://10.217.0.223:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:36:43 crc kubenswrapper[4778]: I0312 13:36:43.936484 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-cell1-metadata-0" podUID="c289a520-78eb-433f-b7a4-0c03be917c18" containerName="nova-cell1-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.223:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 13:36:44 crc kubenswrapper[4778]: I0312 13:36:44.248309 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-1" Mar 12 13:36:44 crc kubenswrapper[4778]: I0312 13:36:44.248607 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-1" Mar 12 13:36:46 crc kubenswrapper[4778]: I0312 13:36:46.267366 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-1" Mar 12 13:36:46 crc kubenswrapper[4778]: I0312 13:36:46.267481 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-1" Mar 12 13:36:46 crc kubenswrapper[4778]: I0312 13:36:46.275831 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-1" Mar 12 13:36:46 crc kubenswrapper[4778]: I0312 13:36:46.277234 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-1" Mar 12 13:36:47 crc kubenswrapper[4778]: I0312 13:36:47.010016 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 12 13:36:47 crc kubenswrapper[4778]: I0312 13:36:47.888607 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 13:36:47 crc kubenswrapper[4778]: I0312 13:36:47.888903 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 13:36:49 crc kubenswrapper[4778]: I0312 13:36:49.895691 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 13:36:49 crc kubenswrapper[4778]: I0312 13:36:49.898408 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 13:36:49 crc kubenswrapper[4778]: I0312 13:36:49.913775 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 13:36:50 crc kubenswrapper[4778]: I0312 13:36:50.479647 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 13:36:50 crc kubenswrapper[4778]: I0312 13:36:50.917361 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:50 crc kubenswrapper[4778]: I0312 13:36:50.917676 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:52 crc kubenswrapper[4778]: I0312 13:36:52.923494 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:52 crc kubenswrapper[4778]: I0312 13:36:52.924209 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:52 crc kubenswrapper[4778]: I0312 13:36:52.928206 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:53 crc kubenswrapper[4778]: I0312 13:36:53.254338 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:36:53 crc kubenswrapper[4778]: E0312 13:36:53.254547 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:36:53 crc kubenswrapper[4778]: I0312 13:36:53.497409 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-metadata-0" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.612551 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.616501 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-566c4d5fc-zx97x" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.779935 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-769c65dfd5-frvxx"] Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.863233 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-566c4d5fc-dggmh"] Mar 12 13:36:54 crc kubenswrapper[4778]: E0312 13:36:54.863646 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="init" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.863662 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="init" Mar 12 13:36:54 crc kubenswrapper[4778]: E0312 13:36:54.863679 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="dnsmasq-dns" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.863686 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="dnsmasq-dns" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.863880 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d621990b-b3fb-457c-a7b8-0726fa89a5e6" containerName="dnsmasq-dns" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.864843 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.874208 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-566c4d5fc-dggmh"] Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.937737 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-public-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.938229 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kpcb\" (UniqueName: \"kubernetes.io/projected/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-kube-api-access-8kpcb\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.938271 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-ovndb-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.938402 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-config\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.938518 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-httpd-config\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.938607 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-combined-ca-bundle\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:54 crc kubenswrapper[4778]: I0312 13:36:54.938826 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-internal-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040630 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-public-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040694 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kpcb\" (UniqueName: \"kubernetes.io/projected/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-kube-api-access-8kpcb\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040725 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-ovndb-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040786 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-config\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040856 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-httpd-config\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040906 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-combined-ca-bundle\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.040936 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-internal-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.056473 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-combined-ca-bundle\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.057021 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-config\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.060462 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-httpd-config\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.060745 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-ovndb-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.061537 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-internal-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.064025 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kpcb\" (UniqueName: \"kubernetes.io/projected/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-kube-api-access-8kpcb\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.064743 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7596a69e-33c9-4a2b-89fc-e4c41252b3fd-public-tls-certs\") pod \"neutron-566c4d5fc-dggmh\" (UID: \"7596a69e-33c9-4a2b-89fc-e4c41252b3fd\") " pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.382865 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.543478 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-769c65dfd5-frvxx" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-api" containerID="cri-o://eb476b810cfe28d9d73622ddd41bf8c8fc415530e6ad67a1faffa32c9bd043ba" gracePeriod=30 Mar 12 13:36:55 crc kubenswrapper[4778]: I0312 13:36:55.544872 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-769c65dfd5-frvxx" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-httpd" containerID="cri-o://1e1f7bdc1b1f277f7f73a02d00233e5689179238b97d5a12c4ec486b0a81ef94" gracePeriod=30 Mar 12 13:36:56 crc kubenswrapper[4778]: I0312 13:36:56.093484 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-566c4d5fc-dggmh"] Mar 12 13:36:56 crc kubenswrapper[4778]: W0312 13:36:56.096057 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7596a69e_33c9_4a2b_89fc_e4c41252b3fd.slice/crio-2eb6ace63a707a07b12839eb74da3ba12196b862b500909424af3b42430d3b71 WatchSource:0}: Error finding container 2eb6ace63a707a07b12839eb74da3ba12196b862b500909424af3b42430d3b71: Status 404 returned error can't find the container with id 2eb6ace63a707a07b12839eb74da3ba12196b862b500909424af3b42430d3b71 Mar 12 13:36:56 crc kubenswrapper[4778]: I0312 13:36:56.698521 4778 generic.go:334] "Generic (PLEG): container finished" podID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerID="1e1f7bdc1b1f277f7f73a02d00233e5689179238b97d5a12c4ec486b0a81ef94" exitCode=0 Mar 12 13:36:56 crc kubenswrapper[4778]: I0312 13:36:56.698605 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-frvxx" event={"ID":"2e1d8894-7234-40d0-b42a-9d7ab1ce638a","Type":"ContainerDied","Data":"1e1f7bdc1b1f277f7f73a02d00233e5689179238b97d5a12c4ec486b0a81ef94"} Mar 12 13:36:56 crc kubenswrapper[4778]: I0312 13:36:56.701487 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c4d5fc-dggmh" event={"ID":"7596a69e-33c9-4a2b-89fc-e4c41252b3fd","Type":"ContainerStarted","Data":"95d8fe1929b2a2cfa341a1e79e3ce52d2a315d8b00026b9096e54bcf440b46e1"} Mar 12 13:36:56 crc kubenswrapper[4778]: I0312 13:36:56.701526 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c4d5fc-dggmh" event={"ID":"7596a69e-33c9-4a2b-89fc-e4c41252b3fd","Type":"ContainerStarted","Data":"2eb6ace63a707a07b12839eb74da3ba12196b862b500909424af3b42430d3b71"} Mar 12 13:36:57 crc kubenswrapper[4778]: I0312 13:36:57.097260 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-69b6dc4885-z4h9m" Mar 12 13:36:57 crc kubenswrapper[4778]: I0312 13:36:57.711443 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c4d5fc-dggmh" event={"ID":"7596a69e-33c9-4a2b-89fc-e4c41252b3fd","Type":"ContainerStarted","Data":"1517d69a357a59e29c554dbc1da03937cc2b1241837601f9ab3357c44fb10e2f"} Mar 12 13:36:57 crc kubenswrapper[4778]: I0312 13:36:57.712147 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:36:57 crc kubenswrapper[4778]: I0312 13:36:57.730201 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-566c4d5fc-dggmh" podStartSLOduration=3.73016016 podStartE2EDuration="3.73016016s" podCreationTimestamp="2026-03-12 13:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:36:57.725858837 +0000 UTC m=+1636.174554253" watchObservedRunningTime="2026-03-12 13:36:57.73016016 +0000 UTC m=+1636.178855576" Mar 12 13:37:01 crc kubenswrapper[4778]: I0312 13:37:01.687622 4778 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-pn8tk" podUID="5e38a4fd-95f8-437b-923b-eca33b1387e6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.70:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 13:37:04 crc kubenswrapper[4778]: I0312 13:37:04.915612 4778 generic.go:334] "Generic (PLEG): container finished" podID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerID="eb476b810cfe28d9d73622ddd41bf8c8fc415530e6ad67a1faffa32c9bd043ba" exitCode=0 Mar 12 13:37:04 crc kubenswrapper[4778]: I0312 13:37:04.915658 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-frvxx" event={"ID":"2e1d8894-7234-40d0-b42a-9d7ab1ce638a","Type":"ContainerDied","Data":"eb476b810cfe28d9d73622ddd41bf8c8fc415530e6ad67a1faffa32c9bd043ba"} Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.320845 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.385575 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-config\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.385697 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-httpd-config\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.385757 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-internal-tls-certs\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.385811 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-public-tls-certs\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.385927 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqjss\" (UniqueName: \"kubernetes.io/projected/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-kube-api-access-nqjss\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.385973 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-combined-ca-bundle\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.386092 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-ovndb-tls-certs\") pod \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\" (UID: \"2e1d8894-7234-40d0-b42a-9d7ab1ce638a\") " Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.392436 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.392510 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-kube-api-access-nqjss" (OuterVolumeSpecName: "kube-api-access-nqjss") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "kube-api-access-nqjss". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.443295 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.445082 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-config" (OuterVolumeSpecName: "config") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.447615 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.454524 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.468336 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2e1d8894-7234-40d0-b42a-9d7ab1ce638a" (UID: "2e1d8894-7234-40d0-b42a-9d7ab1ce638a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.488743 4778 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.488940 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.489029 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.489108 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.489217 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.489307 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqjss\" (UniqueName: \"kubernetes.io/projected/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-kube-api-access-nqjss\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.489372 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1d8894-7234-40d0-b42a-9d7ab1ce638a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.929409 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-frvxx" event={"ID":"2e1d8894-7234-40d0-b42a-9d7ab1ce638a","Type":"ContainerDied","Data":"e2d775202948449d32b3e6f8c66299f17943aeca0f3f57c7b82f6f8283ff7095"} Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.929459 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-frvxx" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.929745 4778 scope.go:117] "RemoveContainer" containerID="1e1f7bdc1b1f277f7f73a02d00233e5689179238b97d5a12c4ec486b0a81ef94" Mar 12 13:37:05 crc kubenswrapper[4778]: I0312 13:37:05.975133 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-769c65dfd5-frvxx"] Mar 12 13:37:06 crc kubenswrapper[4778]: I0312 13:37:06.361129 4778 scope.go:117] "RemoveContainer" containerID="eb476b810cfe28d9d73622ddd41bf8c8fc415530e6ad67a1faffa32c9bd043ba" Mar 12 13:37:06 crc kubenswrapper[4778]: I0312 13:37:06.368714 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-769c65dfd5-frvxx"] Mar 12 13:37:07 crc kubenswrapper[4778]: I0312 13:37:07.254702 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:37:07 crc kubenswrapper[4778]: E0312 13:37:07.255755 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:37:08 crc kubenswrapper[4778]: I0312 13:37:08.267922 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" path="/var/lib/kubelet/pods/2e1d8894-7234-40d0-b42a-9d7ab1ce638a/volumes" Mar 12 13:37:15 crc kubenswrapper[4778]: I0312 13:37:15.909044 4778 scope.go:117] "RemoveContainer" containerID="6addcbc9f6e1bd0c36c2127749a9343943bce9503688868083bfb8596a8eda94" Mar 12 13:37:18 crc kubenswrapper[4778]: I0312 13:37:18.254656 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:37:18 crc kubenswrapper[4778]: E0312 13:37:18.255457 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:37:25 crc kubenswrapper[4778]: I0312 13:37:25.395514 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-566c4d5fc-dggmh" Mar 12 13:37:25 crc kubenswrapper[4778]: I0312 13:37:25.465232 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-769c65dfd5-t7d9g"] Mar 12 13:37:25 crc kubenswrapper[4778]: I0312 13:37:25.465522 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-769c65dfd5-t7d9g" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-api" containerID="cri-o://7559ac32cffd7eca339ac8d8d2f5491100a0167d9ce788c2eb95e805cc071cda" gracePeriod=30 Mar 12 13:37:25 crc kubenswrapper[4778]: I0312 13:37:25.465649 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-769c65dfd5-t7d9g" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-httpd" containerID="cri-o://6738f9dd946d748869f4b26f4030a90ea55b7a4599f29ac178ad859657a706f7" gracePeriod=30 Mar 12 13:37:26 crc kubenswrapper[4778]: I0312 13:37:26.134630 4778 generic.go:334] "Generic (PLEG): container finished" podID="e3118f8b-6bd2-4fba-8300-114513770916" containerID="6738f9dd946d748869f4b26f4030a90ea55b7a4599f29ac178ad859657a706f7" exitCode=0 Mar 12 13:37:26 crc kubenswrapper[4778]: I0312 13:37:26.134695 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-t7d9g" event={"ID":"e3118f8b-6bd2-4fba-8300-114513770916","Type":"ContainerDied","Data":"6738f9dd946d748869f4b26f4030a90ea55b7a4599f29ac178ad859657a706f7"} Mar 12 13:37:32 crc kubenswrapper[4778]: I0312 13:37:32.273668 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:37:32 crc kubenswrapper[4778]: E0312 13:37:32.274437 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.391541 4778 generic.go:334] "Generic (PLEG): container finished" podID="e3118f8b-6bd2-4fba-8300-114513770916" containerID="7559ac32cffd7eca339ac8d8d2f5491100a0167d9ce788c2eb95e805cc071cda" exitCode=0 Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.391616 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-t7d9g" event={"ID":"e3118f8b-6bd2-4fba-8300-114513770916","Type":"ContainerDied","Data":"7559ac32cffd7eca339ac8d8d2f5491100a0167d9ce788c2eb95e805cc071cda"} Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.808566 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.896706 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-combined-ca-bundle\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.896757 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f76t\" (UniqueName: \"kubernetes.io/projected/e3118f8b-6bd2-4fba-8300-114513770916-kube-api-access-4f76t\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.896817 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.896910 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-httpd-config\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.896979 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-public-tls-certs\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.897067 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-internal-tls-certs\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.897206 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-config\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.904246 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.905305 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3118f8b-6bd2-4fba-8300-114513770916-kube-api-access-4f76t" (OuterVolumeSpecName: "kube-api-access-4f76t") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "kube-api-access-4f76t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.948230 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.958733 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.962354 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.969983 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-config" (OuterVolumeSpecName: "config") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.999302 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.999529 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs\") pod \"e3118f8b-6bd2-4fba-8300-114513770916\" (UID: \"e3118f8b-6bd2-4fba-8300-114513770916\") " Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.999957 4778 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:43 crc kubenswrapper[4778]: I0312 13:37:43.999977 4778 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:43.999989 4778 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.000001 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.000010 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.000020 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f76t\" (UniqueName: \"kubernetes.io/projected/e3118f8b-6bd2-4fba-8300-114513770916-kube-api-access-4f76t\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:44 crc kubenswrapper[4778]: W0312 13:37:44.000093 4778 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e3118f8b-6bd2-4fba-8300-114513770916/volumes/kubernetes.io~secret/ovndb-tls-certs Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.000105 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e3118f8b-6bd2-4fba-8300-114513770916" (UID: "e3118f8b-6bd2-4fba-8300-114513770916"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.101351 4778 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3118f8b-6bd2-4fba-8300-114513770916-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.401857 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769c65dfd5-t7d9g" event={"ID":"e3118f8b-6bd2-4fba-8300-114513770916","Type":"ContainerDied","Data":"8300c5c0870d3a0dc15fa6bca84b387efeba0222c0e9b918971777a65c2fcb29"} Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.401911 4778 scope.go:117] "RemoveContainer" containerID="6738f9dd946d748869f4b26f4030a90ea55b7a4599f29ac178ad859657a706f7" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.401920 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769c65dfd5-t7d9g" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.439684 4778 scope.go:117] "RemoveContainer" containerID="7559ac32cffd7eca339ac8d8d2f5491100a0167d9ce788c2eb95e805cc071cda" Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.444158 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-769c65dfd5-t7d9g"] Mar 12 13:37:44 crc kubenswrapper[4778]: I0312 13:37:44.453647 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-769c65dfd5-t7d9g"] Mar 12 13:37:46 crc kubenswrapper[4778]: I0312 13:37:46.254668 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:37:46 crc kubenswrapper[4778]: E0312 13:37:46.255332 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:37:46 crc kubenswrapper[4778]: I0312 13:37:46.265873 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3118f8b-6bd2-4fba-8300-114513770916" path="/var/lib/kubelet/pods/e3118f8b-6bd2-4fba-8300-114513770916/volumes" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.563697 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c5988475-bw257"] Mar 12 13:37:47 crc kubenswrapper[4778]: E0312 13:37:47.564491 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-httpd" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564509 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-httpd" Mar 12 13:37:47 crc kubenswrapper[4778]: E0312 13:37:47.564526 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-api" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564535 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-api" Mar 12 13:37:47 crc kubenswrapper[4778]: E0312 13:37:47.564562 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-api" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564570 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-api" Mar 12 13:37:47 crc kubenswrapper[4778]: E0312 13:37:47.564581 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-httpd" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564588 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-httpd" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564797 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-api" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564827 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-httpd" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564839 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1d8894-7234-40d0-b42a-9d7ab1ce638a" containerName="neutron-api" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.564855 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3118f8b-6bd2-4fba-8300-114513770916" containerName="neutron-httpd" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.565973 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.568058 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.583273 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c5988475-bw257"] Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672036 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-openstack-edpm-ipam\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672092 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672136 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672312 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbdfk\" (UniqueName: \"kubernetes.io/projected/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-kube-api-access-cbdfk\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672448 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-svc\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672508 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-config\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.672617 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.774741 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbdfk\" (UniqueName: \"kubernetes.io/projected/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-kube-api-access-cbdfk\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.774822 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-svc\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.774862 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-config\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776027 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-svc\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776060 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776085 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-config\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776135 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776261 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-openstack-edpm-ipam\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776917 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-openstack-edpm-ipam\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.776986 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.777708 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.778482 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.777797 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.795712 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbdfk\" (UniqueName: \"kubernetes.io/projected/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-kube-api-access-cbdfk\") pod \"dnsmasq-dns-7c5988475-bw257\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:47 crc kubenswrapper[4778]: I0312 13:37:47.886703 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:48 crc kubenswrapper[4778]: I0312 13:37:48.350342 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c5988475-bw257"] Mar 12 13:37:48 crc kubenswrapper[4778]: I0312 13:37:48.466068 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5988475-bw257" event={"ID":"8ebe6a74-d22a-427c-b7a5-f4212457f7d3","Type":"ContainerStarted","Data":"c86a22334e8cef280e44974db1684c5e3e847838867268917e47cc90abe99154"} Mar 12 13:37:49 crc kubenswrapper[4778]: I0312 13:37:49.476066 4778 generic.go:334] "Generic (PLEG): container finished" podID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerID="8675bbbc50526e6bd6a4f0eaa32cd9b3df1cfc38839bff57f7cd99b7bd0da73b" exitCode=0 Mar 12 13:37:49 crc kubenswrapper[4778]: I0312 13:37:49.476146 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5988475-bw257" event={"ID":"8ebe6a74-d22a-427c-b7a5-f4212457f7d3","Type":"ContainerDied","Data":"8675bbbc50526e6bd6a4f0eaa32cd9b3df1cfc38839bff57f7cd99b7bd0da73b"} Mar 12 13:37:50 crc kubenswrapper[4778]: I0312 13:37:50.486872 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5988475-bw257" event={"ID":"8ebe6a74-d22a-427c-b7a5-f4212457f7d3","Type":"ContainerStarted","Data":"cb0a2bf0ec904178e0889259a4ca3a676df6891d1b234fde05f9bd7a8c828b69"} Mar 12 13:37:50 crc kubenswrapper[4778]: I0312 13:37:50.487055 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:57 crc kubenswrapper[4778]: I0312 13:37:57.888448 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:37:57 crc kubenswrapper[4778]: I0312 13:37:57.919501 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c5988475-bw257" podStartSLOduration=10.919477663 podStartE2EDuration="10.919477663s" podCreationTimestamp="2026-03-12 13:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:37:50.505256141 +0000 UTC m=+1688.953951547" watchObservedRunningTime="2026-03-12 13:37:57.919477663 +0000 UTC m=+1696.368173059" Mar 12 13:37:57 crc kubenswrapper[4778]: I0312 13:37:57.967932 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f59c7d6f9-7f6bj"] Mar 12 13:37:57 crc kubenswrapper[4778]: I0312 13:37:57.968286 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerName="dnsmasq-dns" containerID="cri-o://80b9a94e51ace133a39bb4f360454c37e2be50602309d428d0792de3b24d2efc" gracePeriod=10 Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.114040 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f89cfcd7f-vk6h4"] Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.115802 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.194486 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f89cfcd7f-vk6h4"] Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210399 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-dns-svc\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210719 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2xdg\" (UniqueName: \"kubernetes.io/projected/46f34397-57fe-425d-b69d-040f4384ac69-kube-api-access-d2xdg\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210788 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-ovsdbserver-sb\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210832 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210886 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-dns-swift-storage-0\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210922 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-config\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.210966 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-ovsdbserver-nb\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313075 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-ovsdbserver-sb\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313173 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313261 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-dns-swift-storage-0\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313297 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-config\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313371 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-ovsdbserver-nb\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313470 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-dns-svc\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.313512 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2xdg\" (UniqueName: \"kubernetes.io/projected/46f34397-57fe-425d-b69d-040f4384ac69-kube-api-access-d2xdg\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.317859 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-ovsdbserver-sb\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.318846 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.319461 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-dns-swift-storage-0\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.319880 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-ovsdbserver-nb\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.322717 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-config\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.322886 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f34397-57fe-425d-b69d-040f4384ac69-dns-svc\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.338268 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2xdg\" (UniqueName: \"kubernetes.io/projected/46f34397-57fe-425d-b69d-040f4384ac69-kube-api-access-d2xdg\") pod \"dnsmasq-dns-6f89cfcd7f-vk6h4\" (UID: \"46f34397-57fe-425d-b69d-040f4384ac69\") " pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.485877 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.569826 4778 generic.go:334] "Generic (PLEG): container finished" podID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerID="80b9a94e51ace133a39bb4f360454c37e2be50602309d428d0792de3b24d2efc" exitCode=0 Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.569872 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" event={"ID":"a677b5ba-f5d3-4310-ab6d-af0505e82a00","Type":"ContainerDied","Data":"80b9a94e51ace133a39bb4f360454c37e2be50602309d428d0792de3b24d2efc"} Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.569902 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" event={"ID":"a677b5ba-f5d3-4310-ab6d-af0505e82a00","Type":"ContainerDied","Data":"048876f254d8481a39bc4ba587f25ae5e4007ace7976831d743f8095461c0872"} Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.569916 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="048876f254d8481a39bc4ba587f25ae5e4007ace7976831d743f8095461c0872" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.603635 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.630911 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdn6k\" (UniqueName: \"kubernetes.io/projected/a677b5ba-f5d3-4310-ab6d-af0505e82a00-kube-api-access-tdn6k\") pod \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.630984 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-nb\") pod \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.631026 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-swift-storage-0\") pod \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.631295 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-config\") pod \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.631392 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-sb\") pod \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.631436 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-svc\") pod \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\" (UID: \"a677b5ba-f5d3-4310-ab6d-af0505e82a00\") " Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.639340 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a677b5ba-f5d3-4310-ab6d-af0505e82a00-kube-api-access-tdn6k" (OuterVolumeSpecName: "kube-api-access-tdn6k") pod "a677b5ba-f5d3-4310-ab6d-af0505e82a00" (UID: "a677b5ba-f5d3-4310-ab6d-af0505e82a00"). InnerVolumeSpecName "kube-api-access-tdn6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.723961 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a677b5ba-f5d3-4310-ab6d-af0505e82a00" (UID: "a677b5ba-f5d3-4310-ab6d-af0505e82a00"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.724432 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a677b5ba-f5d3-4310-ab6d-af0505e82a00" (UID: "a677b5ba-f5d3-4310-ab6d-af0505e82a00"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.724513 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a677b5ba-f5d3-4310-ab6d-af0505e82a00" (UID: "a677b5ba-f5d3-4310-ab6d-af0505e82a00"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.734536 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-config" (OuterVolumeSpecName: "config") pod "a677b5ba-f5d3-4310-ab6d-af0505e82a00" (UID: "a677b5ba-f5d3-4310-ab6d-af0505e82a00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.734609 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.734635 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdn6k\" (UniqueName: \"kubernetes.io/projected/a677b5ba-f5d3-4310-ab6d-af0505e82a00-kube-api-access-tdn6k\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.734646 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.734658 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.754084 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a677b5ba-f5d3-4310-ab6d-af0505e82a00" (UID: "a677b5ba-f5d3-4310-ab6d-af0505e82a00"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.826597 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f89cfcd7f-vk6h4"] Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.842391 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:58 crc kubenswrapper[4778]: I0312 13:37:58.842441 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a677b5ba-f5d3-4310-ab6d-af0505e82a00-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:37:59 crc kubenswrapper[4778]: I0312 13:37:59.581056 4778 generic.go:334] "Generic (PLEG): container finished" podID="46f34397-57fe-425d-b69d-040f4384ac69" containerID="6513764db666b5964f4fbddc07eb3ed2f92e5d293c8dd60deaf30f6de9a5e9bc" exitCode=0 Mar 12 13:37:59 crc kubenswrapper[4778]: I0312 13:37:59.581133 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" event={"ID":"46f34397-57fe-425d-b69d-040f4384ac69","Type":"ContainerDied","Data":"6513764db666b5964f4fbddc07eb3ed2f92e5d293c8dd60deaf30f6de9a5e9bc"} Mar 12 13:37:59 crc kubenswrapper[4778]: I0312 13:37:59.581519 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" event={"ID":"46f34397-57fe-425d-b69d-040f4384ac69","Type":"ContainerStarted","Data":"e21eb540c79359fc01e4bf7155c7750ecf08139d2b4e8d9fd78a7e1f58ecfaf7"} Mar 12 13:37:59 crc kubenswrapper[4778]: I0312 13:37:59.581559 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f59c7d6f9-7f6bj" Mar 12 13:37:59 crc kubenswrapper[4778]: I0312 13:37:59.824819 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f59c7d6f9-7f6bj"] Mar 12 13:37:59 crc kubenswrapper[4778]: I0312 13:37:59.837395 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f59c7d6f9-7f6bj"] Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.138647 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555378-7skl9"] Mar 12 13:38:00 crc kubenswrapper[4778]: E0312 13:38:00.139652 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerName="dnsmasq-dns" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.139785 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerName="dnsmasq-dns" Mar 12 13:38:00 crc kubenswrapper[4778]: E0312 13:38:00.139867 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerName="init" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.139939 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerName="init" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.140318 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" containerName="dnsmasq-dns" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.141240 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.143943 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.144819 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.150450 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555378-7skl9"] Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.153175 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.265072 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a677b5ba-f5d3-4310-ab6d-af0505e82a00" path="/var/lib/kubelet/pods/a677b5ba-f5d3-4310-ab6d-af0505e82a00/volumes" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.274360 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzvt7\" (UniqueName: \"kubernetes.io/projected/446002fc-0307-4c07-8744-630e76bee9aa-kube-api-access-tzvt7\") pod \"auto-csr-approver-29555378-7skl9\" (UID: \"446002fc-0307-4c07-8744-630e76bee9aa\") " pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.376794 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzvt7\" (UniqueName: \"kubernetes.io/projected/446002fc-0307-4c07-8744-630e76bee9aa-kube-api-access-tzvt7\") pod \"auto-csr-approver-29555378-7skl9\" (UID: \"446002fc-0307-4c07-8744-630e76bee9aa\") " pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.397040 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzvt7\" (UniqueName: \"kubernetes.io/projected/446002fc-0307-4c07-8744-630e76bee9aa-kube-api-access-tzvt7\") pod \"auto-csr-approver-29555378-7skl9\" (UID: \"446002fc-0307-4c07-8744-630e76bee9aa\") " pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.464635 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.592724 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" event={"ID":"46f34397-57fe-425d-b69d-040f4384ac69","Type":"ContainerStarted","Data":"07b29f3ba239b2b280d8de52637eaa5b72eefa66ae849802d64fff11f77e90e9"} Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.592954 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.615243 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" podStartSLOduration=2.61522824 podStartE2EDuration="2.61522824s" podCreationTimestamp="2026-03-12 13:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 13:38:00.60817855 +0000 UTC m=+1699.056873956" watchObservedRunningTime="2026-03-12 13:38:00.61522824 +0000 UTC m=+1699.063923636" Mar 12 13:38:00 crc kubenswrapper[4778]: I0312 13:38:00.919355 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555378-7skl9"] Mar 12 13:38:00 crc kubenswrapper[4778]: W0312 13:38:00.921916 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod446002fc_0307_4c07_8744_630e76bee9aa.slice/crio-73907911d4d2c85fde8f408e872b6090f45f3872f54b62d1d2ba385d502b113d WatchSource:0}: Error finding container 73907911d4d2c85fde8f408e872b6090f45f3872f54b62d1d2ba385d502b113d: Status 404 returned error can't find the container with id 73907911d4d2c85fde8f408e872b6090f45f3872f54b62d1d2ba385d502b113d Mar 12 13:38:01 crc kubenswrapper[4778]: I0312 13:38:01.253375 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:38:01 crc kubenswrapper[4778]: E0312 13:38:01.253955 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:38:01 crc kubenswrapper[4778]: I0312 13:38:01.604553 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555378-7skl9" event={"ID":"446002fc-0307-4c07-8744-630e76bee9aa","Type":"ContainerStarted","Data":"73907911d4d2c85fde8f408e872b6090f45f3872f54b62d1d2ba385d502b113d"} Mar 12 13:38:02 crc kubenswrapper[4778]: I0312 13:38:02.615174 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555378-7skl9" event={"ID":"446002fc-0307-4c07-8744-630e76bee9aa","Type":"ContainerStarted","Data":"bba17f86be2a56502271ccc560c6167ec323fcd74423bccb2a6479d1508bc7e8"} Mar 12 13:38:02 crc kubenswrapper[4778]: I0312 13:38:02.639559 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555378-7skl9" podStartSLOduration=1.520025169 podStartE2EDuration="2.639539472s" podCreationTimestamp="2026-03-12 13:38:00 +0000 UTC" firstStartedPulling="2026-03-12 13:38:00.924328236 +0000 UTC m=+1699.373023632" lastFinishedPulling="2026-03-12 13:38:02.043842499 +0000 UTC m=+1700.492537935" observedRunningTime="2026-03-12 13:38:02.629706172 +0000 UTC m=+1701.078401568" watchObservedRunningTime="2026-03-12 13:38:02.639539472 +0000 UTC m=+1701.088234868" Mar 12 13:38:03 crc kubenswrapper[4778]: I0312 13:38:03.633685 4778 generic.go:334] "Generic (PLEG): container finished" podID="446002fc-0307-4c07-8744-630e76bee9aa" containerID="bba17f86be2a56502271ccc560c6167ec323fcd74423bccb2a6479d1508bc7e8" exitCode=0 Mar 12 13:38:03 crc kubenswrapper[4778]: I0312 13:38:03.633941 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555378-7skl9" event={"ID":"446002fc-0307-4c07-8744-630e76bee9aa","Type":"ContainerDied","Data":"bba17f86be2a56502271ccc560c6167ec323fcd74423bccb2a6479d1508bc7e8"} Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.020908 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.085368 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzvt7\" (UniqueName: \"kubernetes.io/projected/446002fc-0307-4c07-8744-630e76bee9aa-kube-api-access-tzvt7\") pod \"446002fc-0307-4c07-8744-630e76bee9aa\" (UID: \"446002fc-0307-4c07-8744-630e76bee9aa\") " Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.102448 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446002fc-0307-4c07-8744-630e76bee9aa-kube-api-access-tzvt7" (OuterVolumeSpecName: "kube-api-access-tzvt7") pod "446002fc-0307-4c07-8744-630e76bee9aa" (UID: "446002fc-0307-4c07-8744-630e76bee9aa"). InnerVolumeSpecName "kube-api-access-tzvt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.187823 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzvt7\" (UniqueName: \"kubernetes.io/projected/446002fc-0307-4c07-8744-630e76bee9aa-kube-api-access-tzvt7\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.357299 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555372-rddbg"] Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.369904 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555372-rddbg"] Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.654076 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555378-7skl9" event={"ID":"446002fc-0307-4c07-8744-630e76bee9aa","Type":"ContainerDied","Data":"73907911d4d2c85fde8f408e872b6090f45f3872f54b62d1d2ba385d502b113d"} Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.654139 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73907911d4d2c85fde8f408e872b6090f45f3872f54b62d1d2ba385d502b113d" Mar 12 13:38:05 crc kubenswrapper[4778]: I0312 13:38:05.654236 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555378-7skl9" Mar 12 13:38:06 crc kubenswrapper[4778]: I0312 13:38:06.266528 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b7e295-a151-42b0-a8d6-d062d9a42e88" path="/var/lib/kubelet/pods/c0b7e295-a151-42b0-a8d6-d062d9a42e88/volumes" Mar 12 13:38:08 crc kubenswrapper[4778]: I0312 13:38:08.487354 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f89cfcd7f-vk6h4" Mar 12 13:38:08 crc kubenswrapper[4778]: I0312 13:38:08.543031 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c5988475-bw257"] Mar 12 13:38:08 crc kubenswrapper[4778]: I0312 13:38:08.557100 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c5988475-bw257" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerName="dnsmasq-dns" containerID="cri-o://cb0a2bf0ec904178e0889259a4ca3a676df6891d1b234fde05f9bd7a8c828b69" gracePeriod=10 Mar 12 13:38:08 crc kubenswrapper[4778]: I0312 13:38:08.692112 4778 generic.go:334] "Generic (PLEG): container finished" podID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerID="cb0a2bf0ec904178e0889259a4ca3a676df6891d1b234fde05f9bd7a8c828b69" exitCode=0 Mar 12 13:38:08 crc kubenswrapper[4778]: I0312 13:38:08.692175 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5988475-bw257" event={"ID":"8ebe6a74-d22a-427c-b7a5-f4212457f7d3","Type":"ContainerDied","Data":"cb0a2bf0ec904178e0889259a4ca3a676df6891d1b234fde05f9bd7a8c828b69"} Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.059251 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.161073 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-openstack-edpm-ipam\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.161120 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbdfk\" (UniqueName: \"kubernetes.io/projected/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-kube-api-access-cbdfk\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.161244 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-swift-storage-0\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.161336 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-nb\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.161408 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-svc\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.161491 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-sb\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.162017 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-config\") pod \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\" (UID: \"8ebe6a74-d22a-427c-b7a5-f4212457f7d3\") " Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.173648 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-kube-api-access-cbdfk" (OuterVolumeSpecName: "kube-api-access-cbdfk") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "kube-api-access-cbdfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.225905 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.226266 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.228393 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.243645 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.244724 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.254718 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-config" (OuterVolumeSpecName: "config") pod "8ebe6a74-d22a-427c-b7a5-f4212457f7d3" (UID: "8ebe6a74-d22a-427c-b7a5-f4212457f7d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264663 4778 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264711 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264726 4778 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264737 4778 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264750 4778 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-config\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264761 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.264773 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbdfk\" (UniqueName: \"kubernetes.io/projected/8ebe6a74-d22a-427c-b7a5-f4212457f7d3-kube-api-access-cbdfk\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.701934 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5988475-bw257" event={"ID":"8ebe6a74-d22a-427c-b7a5-f4212457f7d3","Type":"ContainerDied","Data":"c86a22334e8cef280e44974db1684c5e3e847838867268917e47cc90abe99154"} Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.701968 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5988475-bw257" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.701994 4778 scope.go:117] "RemoveContainer" containerID="cb0a2bf0ec904178e0889259a4ca3a676df6891d1b234fde05f9bd7a8c828b69" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.744321 4778 scope.go:117] "RemoveContainer" containerID="8675bbbc50526e6bd6a4f0eaa32cd9b3df1cfc38839bff57f7cd99b7bd0da73b" Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.744836 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c5988475-bw257"] Mar 12 13:38:09 crc kubenswrapper[4778]: I0312 13:38:09.753919 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c5988475-bw257"] Mar 12 13:38:10 crc kubenswrapper[4778]: I0312 13:38:10.265038 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" path="/var/lib/kubelet/pods/8ebe6a74-d22a-427c-b7a5-f4212457f7d3/volumes" Mar 12 13:38:15 crc kubenswrapper[4778]: I0312 13:38:15.254413 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:38:15 crc kubenswrapper[4778]: E0312 13:38:15.255215 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:38:16 crc kubenswrapper[4778]: I0312 13:38:16.143167 4778 scope.go:117] "RemoveContainer" containerID="83e30e12aea92ff26adeced3b96dea20e98c42e4bd6fda29118e167bf1eeb711" Mar 12 13:38:16 crc kubenswrapper[4778]: I0312 13:38:16.184218 4778 scope.go:117] "RemoveContainer" containerID="f3f7a33c33e8b6e5c107976dcfe1137727c3f5d14f498dcea6e9df482aee564a" Mar 12 13:38:27 crc kubenswrapper[4778]: I0312 13:38:27.255039 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:38:27 crc kubenswrapper[4778]: E0312 13:38:27.256575 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.770702 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc"] Mar 12 13:38:32 crc kubenswrapper[4778]: E0312 13:38:32.772583 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446002fc-0307-4c07-8744-630e76bee9aa" containerName="oc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.772682 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="446002fc-0307-4c07-8744-630e76bee9aa" containerName="oc" Mar 12 13:38:32 crc kubenswrapper[4778]: E0312 13:38:32.772751 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerName="dnsmasq-dns" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.772813 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerName="dnsmasq-dns" Mar 12 13:38:32 crc kubenswrapper[4778]: E0312 13:38:32.772889 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerName="init" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.772949 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerName="init" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.773218 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="446002fc-0307-4c07-8744-630e76bee9aa" containerName="oc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.773312 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ebe6a74-d22a-427c-b7a5-f4212457f7d3" containerName="dnsmasq-dns" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.774941 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.786234 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.786527 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.786608 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.786787 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.820356 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc"] Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.843554 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qkjn\" (UniqueName: \"kubernetes.io/projected/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-kube-api-access-6qkjn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.843640 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.843673 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.843723 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.945498 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qkjn\" (UniqueName: \"kubernetes.io/projected/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-kube-api-access-6qkjn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.945575 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.945617 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.945670 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.951855 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.952096 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.959843 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:32 crc kubenswrapper[4778]: I0312 13:38:32.963653 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qkjn\" (UniqueName: \"kubernetes.io/projected/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-kube-api-access-6qkjn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:33 crc kubenswrapper[4778]: I0312 13:38:33.112722 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:33 crc kubenswrapper[4778]: I0312 13:38:33.643373 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc"] Mar 12 13:38:34 crc kubenswrapper[4778]: I0312 13:38:34.414838 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" event={"ID":"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb","Type":"ContainerStarted","Data":"033c4c36ae25cc7a6f8501b7708a8e4bd2044e9e6abb0ff12418fb82f4d87df3"} Mar 12 13:38:40 crc kubenswrapper[4778]: I0312 13:38:40.255818 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:38:40 crc kubenswrapper[4778]: E0312 13:38:40.256928 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:38:43 crc kubenswrapper[4778]: I0312 13:38:43.319952 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:38:44 crc kubenswrapper[4778]: I0312 13:38:44.560447 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" event={"ID":"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb","Type":"ContainerStarted","Data":"b669070b5f63e02f0b40f059f58f4f609a161a96501a6bb7535e2eedd63acfe3"} Mar 12 13:38:44 crc kubenswrapper[4778]: I0312 13:38:44.587782 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" podStartSLOduration=2.920033067 podStartE2EDuration="12.587760545s" podCreationTimestamp="2026-03-12 13:38:32 +0000 UTC" firstStartedPulling="2026-03-12 13:38:33.649759869 +0000 UTC m=+1732.098455265" lastFinishedPulling="2026-03-12 13:38:43.317487347 +0000 UTC m=+1741.766182743" observedRunningTime="2026-03-12 13:38:44.578981556 +0000 UTC m=+1743.027676962" watchObservedRunningTime="2026-03-12 13:38:44.587760545 +0000 UTC m=+1743.036455941" Mar 12 13:38:51 crc kubenswrapper[4778]: I0312 13:38:51.254456 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:38:51 crc kubenswrapper[4778]: E0312 13:38:51.255146 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:38:56 crc kubenswrapper[4778]: I0312 13:38:56.680660 4778 generic.go:334] "Generic (PLEG): container finished" podID="bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" containerID="b669070b5f63e02f0b40f059f58f4f609a161a96501a6bb7535e2eedd63acfe3" exitCode=0 Mar 12 13:38:56 crc kubenswrapper[4778]: I0312 13:38:56.681283 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" event={"ID":"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb","Type":"ContainerDied","Data":"b669070b5f63e02f0b40f059f58f4f609a161a96501a6bb7535e2eedd63acfe3"} Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.127979 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.246700 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qkjn\" (UniqueName: \"kubernetes.io/projected/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-kube-api-access-6qkjn\") pod \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.247065 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-inventory\") pod \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.247092 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-ssh-key-openstack-edpm-ipam\") pod \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.247161 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-repo-setup-combined-ca-bundle\") pod \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\" (UID: \"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb\") " Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.252216 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" (UID: "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.253068 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-kube-api-access-6qkjn" (OuterVolumeSpecName: "kube-api-access-6qkjn") pod "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" (UID: "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb"). InnerVolumeSpecName "kube-api-access-6qkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.277762 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" (UID: "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.279734 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-inventory" (OuterVolumeSpecName: "inventory") pod "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" (UID: "bd7ac6b4-5600-45ce-b0ea-199dd4baefcb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.349343 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qkjn\" (UniqueName: \"kubernetes.io/projected/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-kube-api-access-6qkjn\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.349375 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.349386 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.349398 4778 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7ac6b4-5600-45ce-b0ea-199dd4baefcb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.702318 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" event={"ID":"bd7ac6b4-5600-45ce-b0ea-199dd4baefcb","Type":"ContainerDied","Data":"033c4c36ae25cc7a6f8501b7708a8e4bd2044e9e6abb0ff12418fb82f4d87df3"} Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.702377 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="033c4c36ae25cc7a6f8501b7708a8e4bd2044e9e6abb0ff12418fb82f4d87df3" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.702738 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.873532 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx"] Mar 12 13:38:58 crc kubenswrapper[4778]: E0312 13:38:58.873945 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.873965 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.874195 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd7ac6b4-5600-45ce-b0ea-199dd4baefcb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.874847 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.877332 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.878263 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.879107 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.879254 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:38:58 crc kubenswrapper[4778]: I0312 13:38:58.888050 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx"] Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.064109 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.064222 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.064297 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whqnl\" (UniqueName: \"kubernetes.io/projected/b99627a8-43d8-4f7d-90f7-530eda3c2213-kube-api-access-whqnl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.064393 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.165563 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.165637 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.165696 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whqnl\" (UniqueName: \"kubernetes.io/projected/b99627a8-43d8-4f7d-90f7-530eda3c2213-kube-api-access-whqnl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.165795 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.169930 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.170927 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.171009 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.190931 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whqnl\" (UniqueName: \"kubernetes.io/projected/b99627a8-43d8-4f7d-90f7-530eda3c2213-kube-api-access-whqnl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.225483 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:38:59 crc kubenswrapper[4778]: W0312 13:38:59.727565 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb99627a8_43d8_4f7d_90f7_530eda3c2213.slice/crio-dc6fa4d7e880c9ed1330cf9f9750b1850f8f0933c2e607fb21a3cab73809d93c WatchSource:0}: Error finding container dc6fa4d7e880c9ed1330cf9f9750b1850f8f0933c2e607fb21a3cab73809d93c: Status 404 returned error can't find the container with id dc6fa4d7e880c9ed1330cf9f9750b1850f8f0933c2e607fb21a3cab73809d93c Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.727839 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx"] Mar 12 13:38:59 crc kubenswrapper[4778]: I0312 13:38:59.731332 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:39:00 crc kubenswrapper[4778]: I0312 13:39:00.722780 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" event={"ID":"b99627a8-43d8-4f7d-90f7-530eda3c2213","Type":"ContainerStarted","Data":"2be88402a7dbb5865b055bb3ee4db9ccaf014ad6b4e21a2044aee944e26732ea"} Mar 12 13:39:00 crc kubenswrapper[4778]: I0312 13:39:00.723335 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" event={"ID":"b99627a8-43d8-4f7d-90f7-530eda3c2213","Type":"ContainerStarted","Data":"dc6fa4d7e880c9ed1330cf9f9750b1850f8f0933c2e607fb21a3cab73809d93c"} Mar 12 13:39:00 crc kubenswrapper[4778]: I0312 13:39:00.743832 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" podStartSLOduration=2.274380613 podStartE2EDuration="2.743811357s" podCreationTimestamp="2026-03-12 13:38:58 +0000 UTC" firstStartedPulling="2026-03-12 13:38:59.73109828 +0000 UTC m=+1758.179793666" lastFinishedPulling="2026-03-12 13:39:00.200529024 +0000 UTC m=+1758.649224410" observedRunningTime="2026-03-12 13:39:00.739661779 +0000 UTC m=+1759.188357175" watchObservedRunningTime="2026-03-12 13:39:00.743811357 +0000 UTC m=+1759.192506753" Mar 12 13:39:02 crc kubenswrapper[4778]: I0312 13:39:02.254213 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:39:02 crc kubenswrapper[4778]: E0312 13:39:02.254746 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:39:13 crc kubenswrapper[4778]: I0312 13:39:13.253349 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:39:13 crc kubenswrapper[4778]: E0312 13:39:13.254087 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:39:16 crc kubenswrapper[4778]: I0312 13:39:16.320672 4778 scope.go:117] "RemoveContainer" containerID="7c88372c4eebf35fa3a0e19eba355c02e9d34ad468328fc457e997e453d917f3" Mar 12 13:39:16 crc kubenswrapper[4778]: I0312 13:39:16.361508 4778 scope.go:117] "RemoveContainer" containerID="59b401343563918013d35a2531aae9f420a7a4077e0d31999372fd3e7e21e169" Mar 12 13:39:16 crc kubenswrapper[4778]: I0312 13:39:16.432406 4778 scope.go:117] "RemoveContainer" containerID="cc6fc61a82e88c3140b3629f45196f98ee08d5f2fdb0df9b40fe66806a0ccbfd" Mar 12 13:39:28 crc kubenswrapper[4778]: I0312 13:39:28.254720 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:39:28 crc kubenswrapper[4778]: E0312 13:39:28.255519 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:39:42 crc kubenswrapper[4778]: I0312 13:39:42.263822 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:39:42 crc kubenswrapper[4778]: E0312 13:39:42.265702 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:39:53 crc kubenswrapper[4778]: I0312 13:39:53.254204 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:39:53 crc kubenswrapper[4778]: E0312 13:39:53.255018 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.152897 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555380-n8mtp"] Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.154617 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.157011 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.157203 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.157667 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.164770 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555380-n8mtp"] Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.196453 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh2jk\" (UniqueName: \"kubernetes.io/projected/69f54cc7-08e2-42c1-883d-316f1dac7621-kube-api-access-mh2jk\") pod \"auto-csr-approver-29555380-n8mtp\" (UID: \"69f54cc7-08e2-42c1-883d-316f1dac7621\") " pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.298880 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh2jk\" (UniqueName: \"kubernetes.io/projected/69f54cc7-08e2-42c1-883d-316f1dac7621-kube-api-access-mh2jk\") pod \"auto-csr-approver-29555380-n8mtp\" (UID: \"69f54cc7-08e2-42c1-883d-316f1dac7621\") " pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.326260 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh2jk\" (UniqueName: \"kubernetes.io/projected/69f54cc7-08e2-42c1-883d-316f1dac7621-kube-api-access-mh2jk\") pod \"auto-csr-approver-29555380-n8mtp\" (UID: \"69f54cc7-08e2-42c1-883d-316f1dac7621\") " pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.482264 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:00 crc kubenswrapper[4778]: I0312 13:40:00.959688 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555380-n8mtp"] Mar 12 13:40:01 crc kubenswrapper[4778]: I0312 13:40:01.283636 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" event={"ID":"69f54cc7-08e2-42c1-883d-316f1dac7621","Type":"ContainerStarted","Data":"4197c3d5790d7995c1ff07fcc4f70c668110a247508288640324f7eb413c8d5d"} Mar 12 13:40:04 crc kubenswrapper[4778]: I0312 13:40:04.324585 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" event={"ID":"69f54cc7-08e2-42c1-883d-316f1dac7621","Type":"ContainerStarted","Data":"625dea5df6820f4416903072a858eb0ac8d225248a71973001f9856768eaad43"} Mar 12 13:40:04 crc kubenswrapper[4778]: I0312 13:40:04.343600 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" podStartSLOduration=1.5175736149999999 podStartE2EDuration="4.343575935s" podCreationTimestamp="2026-03-12 13:40:00 +0000 UTC" firstStartedPulling="2026-03-12 13:40:00.964767631 +0000 UTC m=+1819.413463027" lastFinishedPulling="2026-03-12 13:40:03.790769951 +0000 UTC m=+1822.239465347" observedRunningTime="2026-03-12 13:40:04.336534065 +0000 UTC m=+1822.785229481" watchObservedRunningTime="2026-03-12 13:40:04.343575935 +0000 UTC m=+1822.792271331" Mar 12 13:40:05 crc kubenswrapper[4778]: I0312 13:40:05.335663 4778 generic.go:334] "Generic (PLEG): container finished" podID="69f54cc7-08e2-42c1-883d-316f1dac7621" containerID="625dea5df6820f4416903072a858eb0ac8d225248a71973001f9856768eaad43" exitCode=0 Mar 12 13:40:05 crc kubenswrapper[4778]: I0312 13:40:05.335723 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" event={"ID":"69f54cc7-08e2-42c1-883d-316f1dac7621","Type":"ContainerDied","Data":"625dea5df6820f4416903072a858eb0ac8d225248a71973001f9856768eaad43"} Mar 12 13:40:06 crc kubenswrapper[4778]: I0312 13:40:06.652799 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:06 crc kubenswrapper[4778]: I0312 13:40:06.733756 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh2jk\" (UniqueName: \"kubernetes.io/projected/69f54cc7-08e2-42c1-883d-316f1dac7621-kube-api-access-mh2jk\") pod \"69f54cc7-08e2-42c1-883d-316f1dac7621\" (UID: \"69f54cc7-08e2-42c1-883d-316f1dac7621\") " Mar 12 13:40:06 crc kubenswrapper[4778]: I0312 13:40:06.739364 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f54cc7-08e2-42c1-883d-316f1dac7621-kube-api-access-mh2jk" (OuterVolumeSpecName: "kube-api-access-mh2jk") pod "69f54cc7-08e2-42c1-883d-316f1dac7621" (UID: "69f54cc7-08e2-42c1-883d-316f1dac7621"). InnerVolumeSpecName "kube-api-access-mh2jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:40:06 crc kubenswrapper[4778]: I0312 13:40:06.837711 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh2jk\" (UniqueName: \"kubernetes.io/projected/69f54cc7-08e2-42c1-883d-316f1dac7621-kube-api-access-mh2jk\") on node \"crc\" DevicePath \"\"" Mar 12 13:40:07 crc kubenswrapper[4778]: I0312 13:40:07.254027 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:40:07 crc kubenswrapper[4778]: E0312 13:40:07.254304 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:40:07 crc kubenswrapper[4778]: I0312 13:40:07.355709 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" event={"ID":"69f54cc7-08e2-42c1-883d-316f1dac7621","Type":"ContainerDied","Data":"4197c3d5790d7995c1ff07fcc4f70c668110a247508288640324f7eb413c8d5d"} Mar 12 13:40:07 crc kubenswrapper[4778]: I0312 13:40:07.355762 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4197c3d5790d7995c1ff07fcc4f70c668110a247508288640324f7eb413c8d5d" Mar 12 13:40:07 crc kubenswrapper[4778]: I0312 13:40:07.355767 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555380-n8mtp" Mar 12 13:40:07 crc kubenswrapper[4778]: I0312 13:40:07.411295 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555374-lf8vj"] Mar 12 13:40:07 crc kubenswrapper[4778]: I0312 13:40:07.419734 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555374-lf8vj"] Mar 12 13:40:08 crc kubenswrapper[4778]: I0312 13:40:08.269634 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d627011-802e-4075-9c56-43373d4c368e" path="/var/lib/kubelet/pods/9d627011-802e-4075-9c56-43373d4c368e/volumes" Mar 12 13:40:16 crc kubenswrapper[4778]: I0312 13:40:16.498257 4778 scope.go:117] "RemoveContainer" containerID="d817d5a09b7856e71332e283d84fe3ea296ae040cb7e986cd73c433864a99c34" Mar 12 13:40:18 crc kubenswrapper[4778]: I0312 13:40:18.253993 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:40:18 crc kubenswrapper[4778]: E0312 13:40:18.254569 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:40:31 crc kubenswrapper[4778]: I0312 13:40:31.254052 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:40:31 crc kubenswrapper[4778]: E0312 13:40:31.254838 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:40:44 crc kubenswrapper[4778]: I0312 13:40:44.254497 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:40:44 crc kubenswrapper[4778]: E0312 13:40:44.255372 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:40:57 crc kubenswrapper[4778]: I0312 13:40:57.254587 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:40:57 crc kubenswrapper[4778]: E0312 13:40:57.255644 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:41:12 crc kubenswrapper[4778]: I0312 13:41:12.260041 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:41:12 crc kubenswrapper[4778]: E0312 13:41:12.260898 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:41:16 crc kubenswrapper[4778]: I0312 13:41:16.577396 4778 scope.go:117] "RemoveContainer" containerID="1e4e2a2aac1ba95c2fc03d3ae5822d197e179d60f0dbd976d4f6143a68eb2c2a" Mar 12 13:41:16 crc kubenswrapper[4778]: I0312 13:41:16.614162 4778 scope.go:117] "RemoveContainer" containerID="1ca532aa466af7c68cb8aa187e7cf3ea161e9610dcf97d902b18dad6b9250f81" Mar 12 13:41:16 crc kubenswrapper[4778]: I0312 13:41:16.662935 4778 scope.go:117] "RemoveContainer" containerID="8dce37445b314b16965ae024d78bbfd9bf5998d5da6305572acf12733671bc3d" Mar 12 13:41:16 crc kubenswrapper[4778]: I0312 13:41:16.704705 4778 scope.go:117] "RemoveContainer" containerID="c66167331bd74d9b577eb48b304f2f99e28d6904a5ae9cd088d4f17df80842e1" Mar 12 13:41:16 crc kubenswrapper[4778]: I0312 13:41:16.728144 4778 scope.go:117] "RemoveContainer" containerID="e6738e925b347d28a1e722ea04cdc7d88018005b75c56a3dec09b214b5752ae1" Mar 12 13:41:23 crc kubenswrapper[4778]: I0312 13:41:23.254630 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:41:23 crc kubenswrapper[4778]: E0312 13:41:23.255528 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:41:30 crc kubenswrapper[4778]: I0312 13:41:30.046269 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-gccjh"] Mar 12 13:41:30 crc kubenswrapper[4778]: I0312 13:41:30.056301 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-gccjh"] Mar 12 13:41:30 crc kubenswrapper[4778]: I0312 13:41:30.263705 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc051b32-4b28-4011-9a00-49caa730f074" path="/var/lib/kubelet/pods/fc051b32-4b28-4011-9a00-49caa730f074/volumes" Mar 12 13:41:32 crc kubenswrapper[4778]: I0312 13:41:32.044407 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-886c-account-create-update-c7kqb"] Mar 12 13:41:32 crc kubenswrapper[4778]: I0312 13:41:32.055129 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-886c-account-create-update-c7kqb"] Mar 12 13:41:32 crc kubenswrapper[4778]: I0312 13:41:32.265473 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b329f80-bb88-4c5c-91eb-24394cdcc492" path="/var/lib/kubelet/pods/7b329f80-bb88-4c5c-91eb-24394cdcc492/volumes" Mar 12 13:41:34 crc kubenswrapper[4778]: I0312 13:41:34.255523 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:41:35 crc kubenswrapper[4778]: I0312 13:41:35.243494 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"92d3dad2e98d7139cb748a76fe93295a7064a4a757626bc932a272018a133968"} Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.040669 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6816-account-create-update-574cj"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.055026 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-79rjc"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.070085 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-hpkvd"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.078107 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-79rjc"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.087359 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6816-account-create-update-574cj"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.096310 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-hpkvd"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.106014 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-3148-account-create-update-zkztc"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.126348 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-3148-account-create-update-zkztc"] Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.276170 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18cd7d9a-1f17-4797-a94f-4692b1180508" path="/var/lib/kubelet/pods/18cd7d9a-1f17-4797-a94f-4692b1180508/volumes" Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.277457 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="280f8bcd-f8e0-451d-8c9c-b733f2b62a23" path="/var/lib/kubelet/pods/280f8bcd-f8e0-451d-8c9c-b733f2b62a23/volumes" Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.278344 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d015b15d-96d2-4b95-9778-8f4175a840a1" path="/var/lib/kubelet/pods/d015b15d-96d2-4b95-9778-8f4175a840a1/volumes" Mar 12 13:41:36 crc kubenswrapper[4778]: I0312 13:41:36.279051 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76971eb-34f0-4a33-b657-508e01eed5d1" path="/var/lib/kubelet/pods/e76971eb-34f0-4a33-b657-508e01eed5d1/volumes" Mar 12 13:41:46 crc kubenswrapper[4778]: I0312 13:41:46.038810 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-7kt6z"] Mar 12 13:41:46 crc kubenswrapper[4778]: I0312 13:41:46.049435 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-7kt6z"] Mar 12 13:41:46 crc kubenswrapper[4778]: I0312 13:41:46.270141 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd5a0cd9-113c-4313-8d66-90487bd90cd3" path="/var/lib/kubelet/pods/dd5a0cd9-113c-4313-8d66-90487bd90cd3/volumes" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.149501 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555382-zbkfk"] Mar 12 13:42:00 crc kubenswrapper[4778]: E0312 13:42:00.150782 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f54cc7-08e2-42c1-883d-316f1dac7621" containerName="oc" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.150800 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f54cc7-08e2-42c1-883d-316f1dac7621" containerName="oc" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.151115 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f54cc7-08e2-42c1-883d-316f1dac7621" containerName="oc" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.152036 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.154844 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.154900 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.154848 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.161861 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555382-zbkfk"] Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.250643 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqvg4\" (UniqueName: \"kubernetes.io/projected/832c789c-468c-400b-8d55-3072443e85ec-kube-api-access-hqvg4\") pod \"auto-csr-approver-29555382-zbkfk\" (UID: \"832c789c-468c-400b-8d55-3072443e85ec\") " pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.352580 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqvg4\" (UniqueName: \"kubernetes.io/projected/832c789c-468c-400b-8d55-3072443e85ec-kube-api-access-hqvg4\") pod \"auto-csr-approver-29555382-zbkfk\" (UID: \"832c789c-468c-400b-8d55-3072443e85ec\") " pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.372025 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqvg4\" (UniqueName: \"kubernetes.io/projected/832c789c-468c-400b-8d55-3072443e85ec-kube-api-access-hqvg4\") pod \"auto-csr-approver-29555382-zbkfk\" (UID: \"832c789c-468c-400b-8d55-3072443e85ec\") " pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.474726 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:00 crc kubenswrapper[4778]: I0312 13:42:00.926542 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555382-zbkfk"] Mar 12 13:42:01 crc kubenswrapper[4778]: I0312 13:42:01.515614 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" event={"ID":"832c789c-468c-400b-8d55-3072443e85ec","Type":"ContainerStarted","Data":"315a9e7ef8d430b73d5cc5b023ba6aa6002c603c03613a82421a60d82ea4b39c"} Mar 12 13:42:03 crc kubenswrapper[4778]: I0312 13:42:03.534152 4778 generic.go:334] "Generic (PLEG): container finished" podID="832c789c-468c-400b-8d55-3072443e85ec" containerID="7785d6a0c6670e984508e3f9d5cc59f211b972f130207a3fed5c63411c140ddc" exitCode=0 Mar 12 13:42:03 crc kubenswrapper[4778]: I0312 13:42:03.534239 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" event={"ID":"832c789c-468c-400b-8d55-3072443e85ec","Type":"ContainerDied","Data":"7785d6a0c6670e984508e3f9d5cc59f211b972f130207a3fed5c63411c140ddc"} Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.204516 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.369393 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqvg4\" (UniqueName: \"kubernetes.io/projected/832c789c-468c-400b-8d55-3072443e85ec-kube-api-access-hqvg4\") pod \"832c789c-468c-400b-8d55-3072443e85ec\" (UID: \"832c789c-468c-400b-8d55-3072443e85ec\") " Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.391594 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/832c789c-468c-400b-8d55-3072443e85ec-kube-api-access-hqvg4" (OuterVolumeSpecName: "kube-api-access-hqvg4") pod "832c789c-468c-400b-8d55-3072443e85ec" (UID: "832c789c-468c-400b-8d55-3072443e85ec"). InnerVolumeSpecName "kube-api-access-hqvg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.473078 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqvg4\" (UniqueName: \"kubernetes.io/projected/832c789c-468c-400b-8d55-3072443e85ec-kube-api-access-hqvg4\") on node \"crc\" DevicePath \"\"" Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.876884 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" event={"ID":"832c789c-468c-400b-8d55-3072443e85ec","Type":"ContainerDied","Data":"315a9e7ef8d430b73d5cc5b023ba6aa6002c603c03613a82421a60d82ea4b39c"} Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.876922 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="315a9e7ef8d430b73d5cc5b023ba6aa6002c603c03613a82421a60d82ea4b39c" Mar 12 13:42:05 crc kubenswrapper[4778]: I0312 13:42:05.876983 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555382-zbkfk" Mar 12 13:42:06 crc kubenswrapper[4778]: I0312 13:42:06.280938 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555376-2bdpv"] Mar 12 13:42:06 crc kubenswrapper[4778]: I0312 13:42:06.288654 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555376-2bdpv"] Mar 12 13:42:08 crc kubenswrapper[4778]: I0312 13:42:08.268339 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f74db3c-fec4-452d-bfd6-8db9f766e0bc" path="/var/lib/kubelet/pods/1f74db3c-fec4-452d-bfd6-8db9f766e0bc/volumes" Mar 12 13:42:11 crc kubenswrapper[4778]: I0312 13:42:11.929303 4778 generic.go:334] "Generic (PLEG): container finished" podID="b99627a8-43d8-4f7d-90f7-530eda3c2213" containerID="2be88402a7dbb5865b055bb3ee4db9ccaf014ad6b4e21a2044aee944e26732ea" exitCode=0 Mar 12 13:42:11 crc kubenswrapper[4778]: I0312 13:42:11.929376 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" event={"ID":"b99627a8-43d8-4f7d-90f7-530eda3c2213","Type":"ContainerDied","Data":"2be88402a7dbb5865b055bb3ee4db9ccaf014ad6b4e21a2044aee944e26732ea"} Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.062070 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-gxsm6"] Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.074697 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-31ed-account-create-update-h8bhm"] Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.087014 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-nh9xs"] Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.099171 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-gxsm6"] Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.111625 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-nh9xs"] Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.126640 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-31ed-account-create-update-h8bhm"] Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.268415 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3800be73-3a09-42b6-8d01-592ccbc6aaa3" path="/var/lib/kubelet/pods/3800be73-3a09-42b6-8d01-592ccbc6aaa3/volumes" Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.269158 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b694c81-3b07-45a1-9ca1-1e47e7430f1f" path="/var/lib/kubelet/pods/4b694c81-3b07-45a1-9ca1-1e47e7430f1f/volumes" Mar 12 13:42:12 crc kubenswrapper[4778]: I0312 13:42:12.269828 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ff3988-1976-4049-8277-0acb36da44c5" path="/var/lib/kubelet/pods/79ff3988-1976-4049-8277-0acb36da44c5/volumes" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.348157 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.451084 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-bootstrap-combined-ca-bundle\") pod \"b99627a8-43d8-4f7d-90f7-530eda3c2213\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.451166 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whqnl\" (UniqueName: \"kubernetes.io/projected/b99627a8-43d8-4f7d-90f7-530eda3c2213-kube-api-access-whqnl\") pod \"b99627a8-43d8-4f7d-90f7-530eda3c2213\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.451319 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-ssh-key-openstack-edpm-ipam\") pod \"b99627a8-43d8-4f7d-90f7-530eda3c2213\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.451355 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-inventory\") pod \"b99627a8-43d8-4f7d-90f7-530eda3c2213\" (UID: \"b99627a8-43d8-4f7d-90f7-530eda3c2213\") " Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.458973 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99627a8-43d8-4f7d-90f7-530eda3c2213-kube-api-access-whqnl" (OuterVolumeSpecName: "kube-api-access-whqnl") pod "b99627a8-43d8-4f7d-90f7-530eda3c2213" (UID: "b99627a8-43d8-4f7d-90f7-530eda3c2213"). InnerVolumeSpecName "kube-api-access-whqnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.459595 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b99627a8-43d8-4f7d-90f7-530eda3c2213" (UID: "b99627a8-43d8-4f7d-90f7-530eda3c2213"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.482452 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-inventory" (OuterVolumeSpecName: "inventory") pod "b99627a8-43d8-4f7d-90f7-530eda3c2213" (UID: "b99627a8-43d8-4f7d-90f7-530eda3c2213"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.485160 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b99627a8-43d8-4f7d-90f7-530eda3c2213" (UID: "b99627a8-43d8-4f7d-90f7-530eda3c2213"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.553770 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.553806 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.553815 4778 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99627a8-43d8-4f7d-90f7-530eda3c2213-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.553823 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whqnl\" (UniqueName: \"kubernetes.io/projected/b99627a8-43d8-4f7d-90f7-530eda3c2213-kube-api-access-whqnl\") on node \"crc\" DevicePath \"\"" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.949531 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" event={"ID":"b99627a8-43d8-4f7d-90f7-530eda3c2213","Type":"ContainerDied","Data":"dc6fa4d7e880c9ed1330cf9f9750b1850f8f0933c2e607fb21a3cab73809d93c"} Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.949575 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx" Mar 12 13:42:13 crc kubenswrapper[4778]: I0312 13:42:13.949580 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc6fa4d7e880c9ed1330cf9f9750b1850f8f0933c2e607fb21a3cab73809d93c" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.062998 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx"] Mar 12 13:42:14 crc kubenswrapper[4778]: E0312 13:42:14.063746 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832c789c-468c-400b-8d55-3072443e85ec" containerName="oc" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.063772 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="832c789c-468c-400b-8d55-3072443e85ec" containerName="oc" Mar 12 13:42:14 crc kubenswrapper[4778]: E0312 13:42:14.063819 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99627a8-43d8-4f7d-90f7-530eda3c2213" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.063831 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99627a8-43d8-4f7d-90f7-530eda3c2213" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.064111 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99627a8-43d8-4f7d-90f7-530eda3c2213" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.064148 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="832c789c-468c-400b-8d55-3072443e85ec" containerName="oc" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.065139 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.070653 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.071827 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.072354 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.072492 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.097818 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx"] Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.166487 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.166938 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgvff\" (UniqueName: \"kubernetes.io/projected/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-kube-api-access-mgvff\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.167259 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.269628 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.269683 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgvff\" (UniqueName: \"kubernetes.io/projected/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-kube-api-access-mgvff\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.269838 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.276006 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.278971 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.291094 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgvff\" (UniqueName: \"kubernetes.io/projected/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-kube-api-access-mgvff\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2xksx\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.397924 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.931806 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx"] Mar 12 13:42:14 crc kubenswrapper[4778]: I0312 13:42:14.969368 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" event={"ID":"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61","Type":"ContainerStarted","Data":"ed030c36635eddb22d286222cfae401d298d4b5f63f4f8d2c4b74f64fe09cc1a"} Mar 12 13:42:15 crc kubenswrapper[4778]: I0312 13:42:15.049516 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-thsh7"] Mar 12 13:42:15 crc kubenswrapper[4778]: I0312 13:42:15.059403 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e25a-account-create-update-vs6zm"] Mar 12 13:42:15 crc kubenswrapper[4778]: I0312 13:42:15.071843 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-thsh7"] Mar 12 13:42:15 crc kubenswrapper[4778]: I0312 13:42:15.081193 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2abd-account-create-update-chtfz"] Mar 12 13:42:15 crc kubenswrapper[4778]: I0312 13:42:15.089550 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e25a-account-create-update-vs6zm"] Mar 12 13:42:15 crc kubenswrapper[4778]: I0312 13:42:15.097439 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2abd-account-create-update-chtfz"] Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.266900 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31f8bb53-a8a8-448f-8f42-349232e383ec" path="/var/lib/kubelet/pods/31f8bb53-a8a8-448f-8f42-349232e383ec/volumes" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.267679 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="729468a8-fded-4564-96c8-471d3cf48825" path="/var/lib/kubelet/pods/729468a8-fded-4564-96c8-471d3cf48825/volumes" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.268323 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9793dfb5-c2a5-4dc1-993d-9e024a810ce8" path="/var/lib/kubelet/pods/9793dfb5-c2a5-4dc1-993d-9e024a810ce8/volumes" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.811339 4778 scope.go:117] "RemoveContainer" containerID="93602c5ae72cfd4f9a42c4921524905037c8077ce8260918d72d9601b072dd59" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.853877 4778 scope.go:117] "RemoveContainer" containerID="70fc2c631648b6cf05ce7c564c8a25d897ce94ea350c4d6a8a0ccacb6c5f16b4" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.894492 4778 scope.go:117] "RemoveContainer" containerID="af7a0409b1470d33d558b70c98a397f0b5c99782ac9578ab1f379f9cb685947f" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.944101 4778 scope.go:117] "RemoveContainer" containerID="d8bdc9c2c4e5e8d5384ac13e3814d6ad0bf996923ba03462051d4c078107d461" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.970722 4778 scope.go:117] "RemoveContainer" containerID="3bf3addaa75cf85838ea1739e9760ca68c0ed5921fd1bd5da9e4725715df9a99" Mar 12 13:42:16 crc kubenswrapper[4778]: I0312 13:42:16.998761 4778 scope.go:117] "RemoveContainer" containerID="2c45c4ddf823adba305999f51111b5e3abaff88105a2366fb93304b13b53f40d" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.020234 4778 scope.go:117] "RemoveContainer" containerID="a62186594073bc08d5194d8b9ce9a46d1a29b359b5ca56b7c0f8fed38f9c7470" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.050472 4778 scope.go:117] "RemoveContainer" containerID="5106184b767437cea31a6a61b3a1991b36587ddd28250ecc1207af703f368fda" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.054928 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" event={"ID":"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61","Type":"ContainerStarted","Data":"07b3a86e7eae89aab9028737c42527be6e9de56b23d4964dfd0e2644aa5dd557"} Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.080030 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" podStartSLOduration=1.952793172 podStartE2EDuration="3.080012558s" podCreationTimestamp="2026-03-12 13:42:14 +0000 UTC" firstStartedPulling="2026-03-12 13:42:14.942436625 +0000 UTC m=+1953.391132021" lastFinishedPulling="2026-03-12 13:42:16.069656011 +0000 UTC m=+1954.518351407" observedRunningTime="2026-03-12 13:42:17.07515684 +0000 UTC m=+1955.523852246" watchObservedRunningTime="2026-03-12 13:42:17.080012558 +0000 UTC m=+1955.528707954" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.092079 4778 scope.go:117] "RemoveContainer" containerID="8443f2894188b4c3d976d78d2d647409527ab07f04b215d8b647fc560059ba2f" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.160892 4778 scope.go:117] "RemoveContainer" containerID="f4bb8c6e00b5e03bcc01c6649d1104fc5ef38426458fa36f98588fb6167dbe07" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.215623 4778 scope.go:117] "RemoveContainer" containerID="13ffa46dd0ede6f8f4fd6e787f1d2948d8a5e96a8e47df52e40147817681f0f7" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.258988 4778 scope.go:117] "RemoveContainer" containerID="60b2242b65665faad21e5afc28edb1788f01dc784524abe26ac1b4cb9a5296a5" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.323002 4778 scope.go:117] "RemoveContainer" containerID="0e8b3287f4617d49763a5e13085485c7f1faa35a7b545d67c3db4b7ac7a3c06b" Mar 12 13:42:17 crc kubenswrapper[4778]: I0312 13:42:17.385568 4778 scope.go:117] "RemoveContainer" containerID="451301ebd2071510b670f3a924d5fcd2f28fbcc4aa60d4224906bca0e09aa5be" Mar 12 13:42:19 crc kubenswrapper[4778]: I0312 13:42:19.029336 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-57cfm"] Mar 12 13:42:19 crc kubenswrapper[4778]: I0312 13:42:19.076500 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-57cfm"] Mar 12 13:42:20 crc kubenswrapper[4778]: I0312 13:42:20.263638 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec77eae6-4dac-4535-b0d3-98bd3422e4de" path="/var/lib/kubelet/pods/ec77eae6-4dac-4535-b0d3-98bd3422e4de/volumes" Mar 12 13:42:28 crc kubenswrapper[4778]: I0312 13:42:28.037667 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-xg6z4"] Mar 12 13:42:28 crc kubenswrapper[4778]: I0312 13:42:28.056696 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-xg6z4"] Mar 12 13:42:28 crc kubenswrapper[4778]: I0312 13:42:28.266619 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="befeb973-a1de-48f9-8de0-5559f75472dc" path="/var/lib/kubelet/pods/befeb973-a1de-48f9-8de0-5559f75472dc/volumes" Mar 12 13:42:51 crc kubenswrapper[4778]: I0312 13:42:51.060669 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-6cvgs"] Mar 12 13:42:51 crc kubenswrapper[4778]: I0312 13:42:51.072465 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zr86r"] Mar 12 13:42:51 crc kubenswrapper[4778]: I0312 13:42:51.083674 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zr86r"] Mar 12 13:42:51 crc kubenswrapper[4778]: I0312 13:42:51.096224 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-6cvgs"] Mar 12 13:42:52 crc kubenswrapper[4778]: I0312 13:42:52.268258 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f8f940-670d-47a0-a90a-afd3aa37a726" path="/var/lib/kubelet/pods/76f8f940-670d-47a0-a90a-afd3aa37a726/volumes" Mar 12 13:42:52 crc kubenswrapper[4778]: I0312 13:42:52.270074 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faeb9cb3-46ae-428f-8c0e-538a2e552072" path="/var/lib/kubelet/pods/faeb9cb3-46ae-428f-8c0e-538a2e552072/volumes" Mar 12 13:42:59 crc kubenswrapper[4778]: I0312 13:42:59.037933 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-56sfj"] Mar 12 13:42:59 crc kubenswrapper[4778]: I0312 13:42:59.047495 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-56sfj"] Mar 12 13:43:00 crc kubenswrapper[4778]: I0312 13:43:00.265394 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af573ef-51c3-4bfc-8de6-eb1be8b75c76" path="/var/lib/kubelet/pods/1af573ef-51c3-4bfc-8de6-eb1be8b75c76/volumes" Mar 12 13:43:04 crc kubenswrapper[4778]: I0312 13:43:04.040989 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-p59s9"] Mar 12 13:43:04 crc kubenswrapper[4778]: I0312 13:43:04.051506 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-p59s9"] Mar 12 13:43:04 crc kubenswrapper[4778]: I0312 13:43:04.266585 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a682334f-73c0-4e38-8f95-e5de661319bb" path="/var/lib/kubelet/pods/a682334f-73c0-4e38-8f95-e5de661319bb/volumes" Mar 12 13:43:13 crc kubenswrapper[4778]: I0312 13:43:13.026243 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-d5pl9"] Mar 12 13:43:13 crc kubenswrapper[4778]: I0312 13:43:13.034449 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-d5pl9"] Mar 12 13:43:14 crc kubenswrapper[4778]: I0312 13:43:14.265576 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb110a1e-6281-437d-b857-eb79c4953e1a" path="/var/lib/kubelet/pods/bb110a1e-6281-437d-b857-eb79c4953e1a/volumes" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.709639 4778 scope.go:117] "RemoveContainer" containerID="fc1fdc3b0586065e85920687a0b5a3f3a3005e79a719fda2a25493dca50c853e" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.747622 4778 scope.go:117] "RemoveContainer" containerID="80b9a94e51ace133a39bb4f360454c37e2be50602309d428d0792de3b24d2efc" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.782486 4778 scope.go:117] "RemoveContainer" containerID="58438369e99b6009fb9ed545548de66fcc857634b3821d960d6e5735646c9d5c" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.818207 4778 scope.go:117] "RemoveContainer" containerID="86b41f2ea1c3794ed3e1fc975ecb18420f64bbd7611743de1aa319532e575758" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.871476 4778 scope.go:117] "RemoveContainer" containerID="434f9dbc426c8bc5145f54de2b34c16cd91006660bd978fe7ad9311fc8579e69" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.912091 4778 scope.go:117] "RemoveContainer" containerID="0a63a4d47752d25e9f6c0d6aa9ed71121a4afe876250e6e10c1c1091bf2b8d8f" Mar 12 13:43:17 crc kubenswrapper[4778]: I0312 13:43:17.957125 4778 scope.go:117] "RemoveContainer" containerID="4711a6f852c8bf6a8fa62e985008d918b7971ec55784fb38d2f086199f1f3aee" Mar 12 13:43:18 crc kubenswrapper[4778]: I0312 13:43:18.008585 4778 scope.go:117] "RemoveContainer" containerID="5a74043e2f16e3024a4f2ed6f0c9502985ad493a8f1362a42f34265b2e50d313" Mar 12 13:43:18 crc kubenswrapper[4778]: I0312 13:43:18.036293 4778 scope.go:117] "RemoveContainer" containerID="710035f2fd1c6ce07427dd61579057ea7d418eb1c9532e9c2ad2d414dc76cbb9" Mar 12 13:43:58 crc kubenswrapper[4778]: I0312 13:43:58.557683 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:43:58 crc kubenswrapper[4778]: I0312 13:43:58.558225 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.145100 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555384-znhr8"] Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.147255 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.150032 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.150071 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.150370 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.160397 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555384-znhr8"] Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.252333 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs6fg\" (UniqueName: \"kubernetes.io/projected/70dc8f5a-da90-4090-b630-a6a7bd438f64-kube-api-access-fs6fg\") pod \"auto-csr-approver-29555384-znhr8\" (UID: \"70dc8f5a-da90-4090-b630-a6a7bd438f64\") " pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.354991 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs6fg\" (UniqueName: \"kubernetes.io/projected/70dc8f5a-da90-4090-b630-a6a7bd438f64-kube-api-access-fs6fg\") pod \"auto-csr-approver-29555384-znhr8\" (UID: \"70dc8f5a-da90-4090-b630-a6a7bd438f64\") " pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.380032 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs6fg\" (UniqueName: \"kubernetes.io/projected/70dc8f5a-da90-4090-b630-a6a7bd438f64-kube-api-access-fs6fg\") pod \"auto-csr-approver-29555384-znhr8\" (UID: \"70dc8f5a-da90-4090-b630-a6a7bd438f64\") " pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.477140 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.967098 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555384-znhr8"] Mar 12 13:44:00 crc kubenswrapper[4778]: I0312 13:44:00.971318 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:44:01 crc kubenswrapper[4778]: I0312 13:44:01.050047 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-sckbb"] Mar 12 13:44:01 crc kubenswrapper[4778]: I0312 13:44:01.060730 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-sckbb"] Mar 12 13:44:01 crc kubenswrapper[4778]: I0312 13:44:01.971629 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555384-znhr8" event={"ID":"70dc8f5a-da90-4090-b630-a6a7bd438f64","Type":"ContainerStarted","Data":"cc8b1411f01c8677b6286c4be7cd084d142c4ecfab7deab72e77cb0b5ddcc743"} Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.049915 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-94ac-account-create-update-rxvgg"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.060438 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2dh9w"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.069168 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-94ac-account-create-update-rxvgg"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.078117 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2dh9w"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.086542 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-dcf9-account-create-update-2rmjd"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.094684 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-x8nht"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.103102 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7d0f-account-create-update-t2rrl"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.111636 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7d0f-account-create-update-t2rrl"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.119480 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-x8nht"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.126335 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-dcf9-account-create-update-2rmjd"] Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.267832 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="068c02bc-1daf-4029-84f9-39a395d5de3e" path="/var/lib/kubelet/pods/068c02bc-1daf-4029-84f9-39a395d5de3e/volumes" Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.269341 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="092c3556-0255-4e2f-b2c7-e22b8a3d8418" path="/var/lib/kubelet/pods/092c3556-0255-4e2f-b2c7-e22b8a3d8418/volumes" Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.270058 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ab681f-51c2-4723-b5b6-58c841185455" path="/var/lib/kubelet/pods/20ab681f-51c2-4723-b5b6-58c841185455/volumes" Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.270756 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d11f6c3-3911-4a29-a65d-ef1f570d9b02" path="/var/lib/kubelet/pods/2d11f6c3-3911-4a29-a65d-ef1f570d9b02/volumes" Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.275900 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad4ff5d-b816-4bdd-97a7-8afd73afe583" path="/var/lib/kubelet/pods/4ad4ff5d-b816-4bdd-97a7-8afd73afe583/volumes" Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.276633 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbbc5fa-b903-4296-a3af-75524920938d" path="/var/lib/kubelet/pods/9dbbc5fa-b903-4296-a3af-75524920938d/volumes" Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.983458 4778 generic.go:334] "Generic (PLEG): container finished" podID="70dc8f5a-da90-4090-b630-a6a7bd438f64" containerID="e97aad250ae3960e7483df5290e0221b9fbbbe6a75ec4afcb92fd5c46ee60b01" exitCode=0 Mar 12 13:44:02 crc kubenswrapper[4778]: I0312 13:44:02.983510 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555384-znhr8" event={"ID":"70dc8f5a-da90-4090-b630-a6a7bd438f64","Type":"ContainerDied","Data":"e97aad250ae3960e7483df5290e0221b9fbbbe6a75ec4afcb92fd5c46ee60b01"} Mar 12 13:44:04 crc kubenswrapper[4778]: I0312 13:44:04.399235 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:04 crc kubenswrapper[4778]: I0312 13:44:04.440004 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs6fg\" (UniqueName: \"kubernetes.io/projected/70dc8f5a-da90-4090-b630-a6a7bd438f64-kube-api-access-fs6fg\") pod \"70dc8f5a-da90-4090-b630-a6a7bd438f64\" (UID: \"70dc8f5a-da90-4090-b630-a6a7bd438f64\") " Mar 12 13:44:04 crc kubenswrapper[4778]: I0312 13:44:04.446451 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70dc8f5a-da90-4090-b630-a6a7bd438f64-kube-api-access-fs6fg" (OuterVolumeSpecName: "kube-api-access-fs6fg") pod "70dc8f5a-da90-4090-b630-a6a7bd438f64" (UID: "70dc8f5a-da90-4090-b630-a6a7bd438f64"). InnerVolumeSpecName "kube-api-access-fs6fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:44:04 crc kubenswrapper[4778]: I0312 13:44:04.542793 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs6fg\" (UniqueName: \"kubernetes.io/projected/70dc8f5a-da90-4090-b630-a6a7bd438f64-kube-api-access-fs6fg\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:05 crc kubenswrapper[4778]: I0312 13:44:05.005673 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555384-znhr8" event={"ID":"70dc8f5a-da90-4090-b630-a6a7bd438f64","Type":"ContainerDied","Data":"cc8b1411f01c8677b6286c4be7cd084d142c4ecfab7deab72e77cb0b5ddcc743"} Mar 12 13:44:05 crc kubenswrapper[4778]: I0312 13:44:05.005713 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555384-znhr8" Mar 12 13:44:05 crc kubenswrapper[4778]: I0312 13:44:05.005730 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc8b1411f01c8677b6286c4be7cd084d142c4ecfab7deab72e77cb0b5ddcc743" Mar 12 13:44:05 crc kubenswrapper[4778]: I0312 13:44:05.457417 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555378-7skl9"] Mar 12 13:44:05 crc kubenswrapper[4778]: I0312 13:44:05.467703 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555378-7skl9"] Mar 12 13:44:06 crc kubenswrapper[4778]: I0312 13:44:06.266015 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="446002fc-0307-4c07-8744-630e76bee9aa" path="/var/lib/kubelet/pods/446002fc-0307-4c07-8744-630e76bee9aa/volumes" Mar 12 13:44:10 crc kubenswrapper[4778]: I0312 13:44:10.043297 4778 generic.go:334] "Generic (PLEG): container finished" podID="96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" containerID="07b3a86e7eae89aab9028737c42527be6e9de56b23d4964dfd0e2644aa5dd557" exitCode=0 Mar 12 13:44:10 crc kubenswrapper[4778]: I0312 13:44:10.043350 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" event={"ID":"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61","Type":"ContainerDied","Data":"07b3a86e7eae89aab9028737c42527be6e9de56b23d4964dfd0e2644aa5dd557"} Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.445142 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.582307 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-inventory\") pod \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.582396 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgvff\" (UniqueName: \"kubernetes.io/projected/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-kube-api-access-mgvff\") pod \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.582528 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-ssh-key-openstack-edpm-ipam\") pod \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\" (UID: \"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61\") " Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.588530 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-kube-api-access-mgvff" (OuterVolumeSpecName: "kube-api-access-mgvff") pod "96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" (UID: "96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61"). InnerVolumeSpecName "kube-api-access-mgvff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.612248 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" (UID: "96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.613362 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-inventory" (OuterVolumeSpecName: "inventory") pod "96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" (UID: "96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.685443 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.685499 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:11 crc kubenswrapper[4778]: I0312 13:44:11.685512 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgvff\" (UniqueName: \"kubernetes.io/projected/96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61-kube-api-access-mgvff\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.062107 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" event={"ID":"96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61","Type":"ContainerDied","Data":"ed030c36635eddb22d286222cfae401d298d4b5f63f4f8d2c4b74f64fe09cc1a"} Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.062523 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed030c36635eddb22d286222cfae401d298d4b5f63f4f8d2c4b74f64fe09cc1a" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.062278 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2xksx" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.174660 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl"] Mar 12 13:44:12 crc kubenswrapper[4778]: E0312 13:44:12.175126 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70dc8f5a-da90-4090-b630-a6a7bd438f64" containerName="oc" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.175150 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="70dc8f5a-da90-4090-b630-a6a7bd438f64" containerName="oc" Mar 12 13:44:12 crc kubenswrapper[4778]: E0312 13:44:12.175201 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.175213 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.175424 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.175463 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="70dc8f5a-da90-4090-b630-a6a7bd438f64" containerName="oc" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.176248 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.178796 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.180626 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.180848 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.182307 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.187693 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl"] Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.301275 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lntxg\" (UniqueName: \"kubernetes.io/projected/5c5541f3-fb44-476b-91c2-b07dffe50894-kube-api-access-lntxg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.301358 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.301415 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.403234 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lntxg\" (UniqueName: \"kubernetes.io/projected/5c5541f3-fb44-476b-91c2-b07dffe50894-kube-api-access-lntxg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.403351 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.403440 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.409474 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.409687 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.424985 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lntxg\" (UniqueName: \"kubernetes.io/projected/5c5541f3-fb44-476b-91c2-b07dffe50894-kube-api-access-lntxg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4szjl\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:12 crc kubenswrapper[4778]: I0312 13:44:12.502316 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:44:13 crc kubenswrapper[4778]: I0312 13:44:13.025521 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl"] Mar 12 13:44:13 crc kubenswrapper[4778]: I0312 13:44:13.076931 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" event={"ID":"5c5541f3-fb44-476b-91c2-b07dffe50894","Type":"ContainerStarted","Data":"2a03fc2efe0354ad242b25364cbd2eba2b14ebdbcdcea201fdb32f5a13b5d430"} Mar 12 13:44:15 crc kubenswrapper[4778]: I0312 13:44:15.096485 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" event={"ID":"5c5541f3-fb44-476b-91c2-b07dffe50894","Type":"ContainerStarted","Data":"f8a3ca31df85cf9da8faf63b19eca49715e39783019bbe431b8fd2f9f7f72be3"} Mar 12 13:44:15 crc kubenswrapper[4778]: I0312 13:44:15.114964 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" podStartSLOduration=1.747757874 podStartE2EDuration="3.114945926s" podCreationTimestamp="2026-03-12 13:44:12 +0000 UTC" firstStartedPulling="2026-03-12 13:44:13.020108487 +0000 UTC m=+2071.468803883" lastFinishedPulling="2026-03-12 13:44:14.387296539 +0000 UTC m=+2072.835991935" observedRunningTime="2026-03-12 13:44:15.112155557 +0000 UTC m=+2073.560850953" watchObservedRunningTime="2026-03-12 13:44:15.114945926 +0000 UTC m=+2073.563641322" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.211846 4778 scope.go:117] "RemoveContainer" containerID="bc734e634b97b1a5646716a6fc635d874255724a3ef890cee0802c7190db7d7c" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.246443 4778 scope.go:117] "RemoveContainer" containerID="bba17f86be2a56502271ccc560c6167ec323fcd74423bccb2a6479d1508bc7e8" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.316776 4778 scope.go:117] "RemoveContainer" containerID="be1aecf0f9c3a392b6320f1bb26caafd070dc71ba9db9be7a31ee5daf79e1a2d" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.382582 4778 scope.go:117] "RemoveContainer" containerID="099862fee239f9d58b6485a586d53c0613281de24cf1629f41917394af426901" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.408616 4778 scope.go:117] "RemoveContainer" containerID="2919ec7bf1dcd65b4aaec3b3c75478bba66c6d492f7b5c0064c9c993485c3e21" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.466577 4778 scope.go:117] "RemoveContainer" containerID="e2a8d1e05ff7ff80a86b71f26e5fb5c7484878b8a9632420829088d85ad0fbaf" Mar 12 13:44:18 crc kubenswrapper[4778]: I0312 13:44:18.528281 4778 scope.go:117] "RemoveContainer" containerID="37e7dd5198914cc6a22b8658dd88edbdbdabb2bfe43c9c4d07a686c73a997ca2" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.660303 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cfmvx"] Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.664589 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.677936 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cfmvx"] Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.755240 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8php4\" (UniqueName: \"kubernetes.io/projected/76d361e0-0808-41e0-a659-d9977bce86de-kube-api-access-8php4\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.755400 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-catalog-content\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.755528 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-utilities\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.858201 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-utilities\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.858289 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8php4\" (UniqueName: \"kubernetes.io/projected/76d361e0-0808-41e0-a659-d9977bce86de-kube-api-access-8php4\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.858428 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-catalog-content\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.858762 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-utilities\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.859136 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-catalog-content\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.878398 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8php4\" (UniqueName: \"kubernetes.io/projected/76d361e0-0808-41e0-a659-d9977bce86de-kube-api-access-8php4\") pod \"certified-operators-cfmvx\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:19 crc kubenswrapper[4778]: I0312 13:44:19.992403 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:20 crc kubenswrapper[4778]: I0312 13:44:20.322359 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cfmvx"] Mar 12 13:44:21 crc kubenswrapper[4778]: I0312 13:44:21.149991 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfmvx" event={"ID":"76d361e0-0808-41e0-a659-d9977bce86de","Type":"ContainerStarted","Data":"af4ac1b2a11604b423a1cb99788d39c58084666f21c326e7b0bc198fdab855e6"} Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.053092 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.055042 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.066516 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.159385 4778 generic.go:334] "Generic (PLEG): container finished" podID="76d361e0-0808-41e0-a659-d9977bce86de" containerID="28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e" exitCode=0 Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.159447 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfmvx" event={"ID":"76d361e0-0808-41e0-a659-d9977bce86de","Type":"ContainerDied","Data":"28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e"} Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.224769 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-utilities\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.224851 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d6fn\" (UniqueName: \"kubernetes.io/projected/c8cc55b1-e6ed-4790-886c-fabe5917bf27-kube-api-access-8d6fn\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.225342 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-catalog-content\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.327322 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-catalog-content\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.327644 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-utilities\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.327674 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d6fn\" (UniqueName: \"kubernetes.io/projected/c8cc55b1-e6ed-4790-886c-fabe5917bf27-kube-api-access-8d6fn\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.327835 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-catalog-content\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.328117 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-utilities\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.356031 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d6fn\" (UniqueName: \"kubernetes.io/projected/c8cc55b1-e6ed-4790-886c-fabe5917bf27-kube-api-access-8d6fn\") pod \"community-operators-zbbfg\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.376270 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.660310 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tsbbw"] Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.664134 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.671813 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsbbw"] Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.843754 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-utilities\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.843998 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wnw4\" (UniqueName: \"kubernetes.io/projected/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-kube-api-access-2wnw4\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.844045 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-catalog-content\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.900975 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.946306 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-utilities\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.946486 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wnw4\" (UniqueName: \"kubernetes.io/projected/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-kube-api-access-2wnw4\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.946531 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-catalog-content\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.947101 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-catalog-content\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.947474 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-utilities\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:22 crc kubenswrapper[4778]: I0312 13:44:22.965304 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wnw4\" (UniqueName: \"kubernetes.io/projected/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-kube-api-access-2wnw4\") pod \"redhat-marketplace-tsbbw\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:23 crc kubenswrapper[4778]: I0312 13:44:23.001167 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:23 crc kubenswrapper[4778]: I0312 13:44:23.187025 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbbfg" event={"ID":"c8cc55b1-e6ed-4790-886c-fabe5917bf27","Type":"ContainerStarted","Data":"7cfbf75bc1bea8190b4fd8a7b4f36c4f8056d3512bf0a0494d17fb32c82abce1"} Mar 12 13:44:23 crc kubenswrapper[4778]: I0312 13:44:23.461543 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsbbw"] Mar 12 13:44:23 crc kubenswrapper[4778]: W0312 13:44:23.465055 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f9b6c4c_ccc9_42ac_9d52_64690d25a4d6.slice/crio-c74f7a8b7ab0e3f37ef2ae936c5f539034e20dd9a81db59b3381c8571e8dbe74 WatchSource:0}: Error finding container c74f7a8b7ab0e3f37ef2ae936c5f539034e20dd9a81db59b3381c8571e8dbe74: Status 404 returned error can't find the container with id c74f7a8b7ab0e3f37ef2ae936c5f539034e20dd9a81db59b3381c8571e8dbe74 Mar 12 13:44:24 crc kubenswrapper[4778]: I0312 13:44:24.206556 4778 generic.go:334] "Generic (PLEG): container finished" podID="76d361e0-0808-41e0-a659-d9977bce86de" containerID="276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690" exitCode=0 Mar 12 13:44:24 crc kubenswrapper[4778]: I0312 13:44:24.206784 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfmvx" event={"ID":"76d361e0-0808-41e0-a659-d9977bce86de","Type":"ContainerDied","Data":"276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690"} Mar 12 13:44:24 crc kubenswrapper[4778]: I0312 13:44:24.209862 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerStarted","Data":"c74f7a8b7ab0e3f37ef2ae936c5f539034e20dd9a81db59b3381c8571e8dbe74"} Mar 12 13:44:24 crc kubenswrapper[4778]: I0312 13:44:24.215972 4778 generic.go:334] "Generic (PLEG): container finished" podID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerID="a44a31875240c27026c8d5b3562efaf0a4ac960ee6a568ff9dac9567138bfecd" exitCode=0 Mar 12 13:44:24 crc kubenswrapper[4778]: I0312 13:44:24.216390 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbbfg" event={"ID":"c8cc55b1-e6ed-4790-886c-fabe5917bf27","Type":"ContainerDied","Data":"a44a31875240c27026c8d5b3562efaf0a4ac960ee6a568ff9dac9567138bfecd"} Mar 12 13:44:25 crc kubenswrapper[4778]: I0312 13:44:25.228947 4778 generic.go:334] "Generic (PLEG): container finished" podID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerID="b80555cb1dcab02445567c3e276216686d7719c0ca0475be3608a7a9c0f26718" exitCode=0 Mar 12 13:44:25 crc kubenswrapper[4778]: I0312 13:44:25.229083 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerDied","Data":"b80555cb1dcab02445567c3e276216686d7719c0ca0475be3608a7a9c0f26718"} Mar 12 13:44:26 crc kubenswrapper[4778]: I0312 13:44:26.239229 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerStarted","Data":"e4ef6232820924fc9bfa9e22782695ff3623174a4e886e71c18dbcccc1dd4b88"} Mar 12 13:44:26 crc kubenswrapper[4778]: I0312 13:44:26.242054 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfmvx" event={"ID":"76d361e0-0808-41e0-a659-d9977bce86de","Type":"ContainerStarted","Data":"a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9"} Mar 12 13:44:26 crc kubenswrapper[4778]: I0312 13:44:26.275962 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cfmvx" podStartSLOduration=4.36772276 podStartE2EDuration="7.275939492s" podCreationTimestamp="2026-03-12 13:44:19 +0000 UTC" firstStartedPulling="2026-03-12 13:44:22.161218631 +0000 UTC m=+2080.609914027" lastFinishedPulling="2026-03-12 13:44:25.069435363 +0000 UTC m=+2083.518130759" observedRunningTime="2026-03-12 13:44:26.273952676 +0000 UTC m=+2084.722648092" watchObservedRunningTime="2026-03-12 13:44:26.275939492 +0000 UTC m=+2084.724634888" Mar 12 13:44:27 crc kubenswrapper[4778]: I0312 13:44:27.253502 4778 generic.go:334] "Generic (PLEG): container finished" podID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerID="e4ef6232820924fc9bfa9e22782695ff3623174a4e886e71c18dbcccc1dd4b88" exitCode=0 Mar 12 13:44:27 crc kubenswrapper[4778]: I0312 13:44:27.253564 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerDied","Data":"e4ef6232820924fc9bfa9e22782695ff3623174a4e886e71c18dbcccc1dd4b88"} Mar 12 13:44:28 crc kubenswrapper[4778]: I0312 13:44:28.557952 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:44:28 crc kubenswrapper[4778]: I0312 13:44:28.558271 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:44:29 crc kubenswrapper[4778]: I0312 13:44:29.994367 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:29 crc kubenswrapper[4778]: I0312 13:44:29.994728 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.037053 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6b6mv"] Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.047497 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6b6mv"] Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.157633 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cfmvx" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="registry-server" probeResult="failure" output=< Mar 12 13:44:31 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:44:31 crc kubenswrapper[4778]: > Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.296083 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerStarted","Data":"53b4b05091c277f65d760aedbdd673f748699fdc3c86d1184ac2f57494d75369"} Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.299557 4778 generic.go:334] "Generic (PLEG): container finished" podID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerID="bdf54c6d37ca16db7981b38aa8bdf481e8ce434ef1861261a6875f0a169c6607" exitCode=0 Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.299652 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbbfg" event={"ID":"c8cc55b1-e6ed-4790-886c-fabe5917bf27","Type":"ContainerDied","Data":"bdf54c6d37ca16db7981b38aa8bdf481e8ce434ef1861261a6875f0a169c6607"} Mar 12 13:44:31 crc kubenswrapper[4778]: I0312 13:44:31.319713 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tsbbw" podStartSLOduration=4.233053285 podStartE2EDuration="9.319694152s" podCreationTimestamp="2026-03-12 13:44:22 +0000 UTC" firstStartedPulling="2026-03-12 13:44:25.234139002 +0000 UTC m=+2083.682834418" lastFinishedPulling="2026-03-12 13:44:30.320779889 +0000 UTC m=+2088.769475285" observedRunningTime="2026-03-12 13:44:31.317011326 +0000 UTC m=+2089.765706732" watchObservedRunningTime="2026-03-12 13:44:31.319694152 +0000 UTC m=+2089.768389548" Mar 12 13:44:32 crc kubenswrapper[4778]: I0312 13:44:32.266367 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe24691f-9019-44ec-85bf-b477c53f05ec" path="/var/lib/kubelet/pods/fe24691f-9019-44ec-85bf-b477c53f05ec/volumes" Mar 12 13:44:32 crc kubenswrapper[4778]: I0312 13:44:32.316569 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbbfg" event={"ID":"c8cc55b1-e6ed-4790-886c-fabe5917bf27","Type":"ContainerStarted","Data":"ed078967e33cb44c74a365a9804f9a8509ee01d3f7a8039f9f7b8f3366ab7aae"} Mar 12 13:44:32 crc kubenswrapper[4778]: I0312 13:44:32.338860 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zbbfg" podStartSLOduration=3.8613881770000003 podStartE2EDuration="10.338837998s" podCreationTimestamp="2026-03-12 13:44:22 +0000 UTC" firstStartedPulling="2026-03-12 13:44:25.231752284 +0000 UTC m=+2083.680447690" lastFinishedPulling="2026-03-12 13:44:31.709202115 +0000 UTC m=+2090.157897511" observedRunningTime="2026-03-12 13:44:32.337675555 +0000 UTC m=+2090.786370951" watchObservedRunningTime="2026-03-12 13:44:32.338837998 +0000 UTC m=+2090.787533394" Mar 12 13:44:32 crc kubenswrapper[4778]: I0312 13:44:32.376817 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:32 crc kubenswrapper[4778]: I0312 13:44:32.376882 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:33 crc kubenswrapper[4778]: I0312 13:44:33.002040 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:33 crc kubenswrapper[4778]: I0312 13:44:33.002290 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:33 crc kubenswrapper[4778]: I0312 13:44:33.425515 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zbbfg" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="registry-server" probeResult="failure" output=< Mar 12 13:44:33 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:44:33 crc kubenswrapper[4778]: > Mar 12 13:44:34 crc kubenswrapper[4778]: I0312 13:44:34.046864 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-tsbbw" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="registry-server" probeResult="failure" output=< Mar 12 13:44:34 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:44:34 crc kubenswrapper[4778]: > Mar 12 13:44:41 crc kubenswrapper[4778]: I0312 13:44:41.067413 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cfmvx" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="registry-server" probeResult="failure" output=< Mar 12 13:44:41 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:44:41 crc kubenswrapper[4778]: > Mar 12 13:44:42 crc kubenswrapper[4778]: I0312 13:44:42.431132 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:42 crc kubenswrapper[4778]: I0312 13:44:42.480366 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 13:44:42 crc kubenswrapper[4778]: I0312 13:44:42.574119 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 13:44:42 crc kubenswrapper[4778]: I0312 13:44:42.670238 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-scbxn"] Mar 12 13:44:42 crc kubenswrapper[4778]: I0312 13:44:42.670544 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-scbxn" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="registry-server" containerID="cri-o://b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76" gracePeriod=2 Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.216699 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.317933 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-catalog-content\") pod \"f2f91915-3841-4662-88e4-82a22df0b131\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.318074 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpxbz\" (UniqueName: \"kubernetes.io/projected/f2f91915-3841-4662-88e4-82a22df0b131-kube-api-access-xpxbz\") pod \"f2f91915-3841-4662-88e4-82a22df0b131\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.318166 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-utilities\") pod \"f2f91915-3841-4662-88e4-82a22df0b131\" (UID: \"f2f91915-3841-4662-88e4-82a22df0b131\") " Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.319495 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-utilities" (OuterVolumeSpecName: "utilities") pod "f2f91915-3841-4662-88e4-82a22df0b131" (UID: "f2f91915-3841-4662-88e4-82a22df0b131"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.339441 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2f91915-3841-4662-88e4-82a22df0b131-kube-api-access-xpxbz" (OuterVolumeSpecName: "kube-api-access-xpxbz") pod "f2f91915-3841-4662-88e4-82a22df0b131" (UID: "f2f91915-3841-4662-88e4-82a22df0b131"). InnerVolumeSpecName "kube-api-access-xpxbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.424446 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpxbz\" (UniqueName: \"kubernetes.io/projected/f2f91915-3841-4662-88e4-82a22df0b131-kube-api-access-xpxbz\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.424489 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.431342 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2f91915-3841-4662-88e4-82a22df0b131" (UID: "f2f91915-3841-4662-88e4-82a22df0b131"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.475396 4778 generic.go:334] "Generic (PLEG): container finished" podID="f2f91915-3841-4662-88e4-82a22df0b131" containerID="b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76" exitCode=0 Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.476310 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-scbxn" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.476550 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerDied","Data":"b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76"} Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.476590 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scbxn" event={"ID":"f2f91915-3841-4662-88e4-82a22df0b131","Type":"ContainerDied","Data":"a19d957b7ce97a07ce6c0132cc4944c7bac635fad459f95aeb77803a9db2f905"} Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.476611 4778 scope.go:117] "RemoveContainer" containerID="b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.513313 4778 scope.go:117] "RemoveContainer" containerID="814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.528466 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f91915-3841-4662-88e4-82a22df0b131-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.547794 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-scbxn"] Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.560857 4778 scope.go:117] "RemoveContainer" containerID="10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.567616 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-scbxn"] Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.628626 4778 scope.go:117] "RemoveContainer" containerID="b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76" Mar 12 13:44:43 crc kubenswrapper[4778]: E0312 13:44:43.629328 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76\": container with ID starting with b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76 not found: ID does not exist" containerID="b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.629372 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76"} err="failed to get container status \"b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76\": rpc error: code = NotFound desc = could not find container \"b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76\": container with ID starting with b5fb35881f73fed40f0d045e20b298d8e653b8e77b271e499bac524ab74b5a76 not found: ID does not exist" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.629401 4778 scope.go:117] "RemoveContainer" containerID="814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e" Mar 12 13:44:43 crc kubenswrapper[4778]: E0312 13:44:43.631416 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e\": container with ID starting with 814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e not found: ID does not exist" containerID="814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.631462 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e"} err="failed to get container status \"814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e\": rpc error: code = NotFound desc = could not find container \"814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e\": container with ID starting with 814289091d5196e8cb90a27db5c8b7b0001e258e89f3f752a355b1aba2fbc07e not found: ID does not exist" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.631495 4778 scope.go:117] "RemoveContainer" containerID="10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62" Mar 12 13:44:43 crc kubenswrapper[4778]: E0312 13:44:43.632342 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62\": container with ID starting with 10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62 not found: ID does not exist" containerID="10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62" Mar 12 13:44:43 crc kubenswrapper[4778]: I0312 13:44:43.632371 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62"} err="failed to get container status \"10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62\": rpc error: code = NotFound desc = could not find container \"10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62\": container with ID starting with 10d3561207dda57207dda1824471a89613d2f3cdc7e885ebdb851a3821e79c62 not found: ID does not exist" Mar 12 13:44:44 crc kubenswrapper[4778]: I0312 13:44:44.145761 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-tsbbw" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="registry-server" probeResult="failure" output=< Mar 12 13:44:44 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:44:44 crc kubenswrapper[4778]: > Mar 12 13:44:44 crc kubenswrapper[4778]: I0312 13:44:44.266767 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2f91915-3841-4662-88e4-82a22df0b131" path="/var/lib/kubelet/pods/f2f91915-3841-4662-88e4-82a22df0b131/volumes" Mar 12 13:44:50 crc kubenswrapper[4778]: I0312 13:44:50.046574 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:50 crc kubenswrapper[4778]: I0312 13:44:50.138794 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:50 crc kubenswrapper[4778]: I0312 13:44:50.859999 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cfmvx"] Mar 12 13:44:51 crc kubenswrapper[4778]: I0312 13:44:51.543156 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cfmvx" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="registry-server" containerID="cri-o://a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9" gracePeriod=2 Mar 12 13:44:51 crc kubenswrapper[4778]: I0312 13:44:51.994760 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.110270 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-catalog-content\") pod \"76d361e0-0808-41e0-a659-d9977bce86de\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.110312 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-utilities\") pod \"76d361e0-0808-41e0-a659-d9977bce86de\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.110441 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8php4\" (UniqueName: \"kubernetes.io/projected/76d361e0-0808-41e0-a659-d9977bce86de-kube-api-access-8php4\") pod \"76d361e0-0808-41e0-a659-d9977bce86de\" (UID: \"76d361e0-0808-41e0-a659-d9977bce86de\") " Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.111297 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-utilities" (OuterVolumeSpecName: "utilities") pod "76d361e0-0808-41e0-a659-d9977bce86de" (UID: "76d361e0-0808-41e0-a659-d9977bce86de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.115693 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d361e0-0808-41e0-a659-d9977bce86de-kube-api-access-8php4" (OuterVolumeSpecName: "kube-api-access-8php4") pod "76d361e0-0808-41e0-a659-d9977bce86de" (UID: "76d361e0-0808-41e0-a659-d9977bce86de"). InnerVolumeSpecName "kube-api-access-8php4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.163849 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76d361e0-0808-41e0-a659-d9977bce86de" (UID: "76d361e0-0808-41e0-a659-d9977bce86de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.212311 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.212353 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d361e0-0808-41e0-a659-d9977bce86de-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.212368 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8php4\" (UniqueName: \"kubernetes.io/projected/76d361e0-0808-41e0-a659-d9977bce86de-kube-api-access-8php4\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.552974 4778 generic.go:334] "Generic (PLEG): container finished" podID="76d361e0-0808-41e0-a659-d9977bce86de" containerID="a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9" exitCode=0 Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.553048 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfmvx" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.553038 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfmvx" event={"ID":"76d361e0-0808-41e0-a659-d9977bce86de","Type":"ContainerDied","Data":"a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9"} Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.553389 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfmvx" event={"ID":"76d361e0-0808-41e0-a659-d9977bce86de","Type":"ContainerDied","Data":"af4ac1b2a11604b423a1cb99788d39c58084666f21c326e7b0bc198fdab855e6"} Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.553407 4778 scope.go:117] "RemoveContainer" containerID="a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.576170 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cfmvx"] Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.580073 4778 scope.go:117] "RemoveContainer" containerID="276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.584825 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cfmvx"] Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.604941 4778 scope.go:117] "RemoveContainer" containerID="28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.646147 4778 scope.go:117] "RemoveContainer" containerID="a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9" Mar 12 13:44:52 crc kubenswrapper[4778]: E0312 13:44:52.646724 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9\": container with ID starting with a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9 not found: ID does not exist" containerID="a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.646770 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9"} err="failed to get container status \"a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9\": rpc error: code = NotFound desc = could not find container \"a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9\": container with ID starting with a4cf397b97c3a80100f006e2a53fcbacdbb24a4168b3b2d40704492acf9072a9 not found: ID does not exist" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.646798 4778 scope.go:117] "RemoveContainer" containerID="276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690" Mar 12 13:44:52 crc kubenswrapper[4778]: E0312 13:44:52.647206 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690\": container with ID starting with 276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690 not found: ID does not exist" containerID="276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.647241 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690"} err="failed to get container status \"276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690\": rpc error: code = NotFound desc = could not find container \"276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690\": container with ID starting with 276ec13767567c5795991c6f0969324f5d851bfcaf87a8fb83fbe5e9808be690 not found: ID does not exist" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.647259 4778 scope.go:117] "RemoveContainer" containerID="28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e" Mar 12 13:44:52 crc kubenswrapper[4778]: E0312 13:44:52.647563 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e\": container with ID starting with 28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e not found: ID does not exist" containerID="28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e" Mar 12 13:44:52 crc kubenswrapper[4778]: I0312 13:44:52.647591 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e"} err="failed to get container status \"28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e\": rpc error: code = NotFound desc = could not find container \"28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e\": container with ID starting with 28bd951139caec3680a15c4aded1c0bca6f1af5a51137c86d0e3f15a3053684e not found: ID does not exist" Mar 12 13:44:53 crc kubenswrapper[4778]: I0312 13:44:53.047613 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:53 crc kubenswrapper[4778]: I0312 13:44:53.093561 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:54 crc kubenswrapper[4778]: I0312 13:44:54.062396 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qqx6r"] Mar 12 13:44:54 crc kubenswrapper[4778]: I0312 13:44:54.079207 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qqx6r"] Mar 12 13:44:54 crc kubenswrapper[4778]: I0312 13:44:54.263787 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d361e0-0808-41e0-a659-d9977bce86de" path="/var/lib/kubelet/pods/76d361e0-0808-41e0-a659-d9977bce86de/volumes" Mar 12 13:44:54 crc kubenswrapper[4778]: I0312 13:44:54.264485 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a74774-1415-43d1-b278-bead87ab4385" path="/var/lib/kubelet/pods/98a74774-1415-43d1-b278-bead87ab4385/volumes" Mar 12 13:44:55 crc kubenswrapper[4778]: I0312 13:44:55.027388 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7dlt6"] Mar 12 13:44:55 crc kubenswrapper[4778]: I0312 13:44:55.037374 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7dlt6"] Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.272484 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58dfb2fb-928e-46de-90dd-481c91a7727c" path="/var/lib/kubelet/pods/58dfb2fb-928e-46de-90dd-481c91a7727c/volumes" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.279983 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsbbw"] Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.280286 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tsbbw" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="registry-server" containerID="cri-o://53b4b05091c277f65d760aedbdd673f748699fdc3c86d1184ac2f57494d75369" gracePeriod=2 Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.596300 4778 generic.go:334] "Generic (PLEG): container finished" podID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerID="53b4b05091c277f65d760aedbdd673f748699fdc3c86d1184ac2f57494d75369" exitCode=0 Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.596349 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerDied","Data":"53b4b05091c277f65d760aedbdd673f748699fdc3c86d1184ac2f57494d75369"} Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.750273 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.800064 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-utilities\") pod \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.800280 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wnw4\" (UniqueName: \"kubernetes.io/projected/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-kube-api-access-2wnw4\") pod \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.800392 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-catalog-content\") pod \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\" (UID: \"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6\") " Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.801074 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-utilities" (OuterVolumeSpecName: "utilities") pod "0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" (UID: "0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.819394 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-kube-api-access-2wnw4" (OuterVolumeSpecName: "kube-api-access-2wnw4") pod "0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" (UID: "0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6"). InnerVolumeSpecName "kube-api-access-2wnw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.829109 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" (UID: "0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.902991 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.903258 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wnw4\" (UniqueName: \"kubernetes.io/projected/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-kube-api-access-2wnw4\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:56 crc kubenswrapper[4778]: I0312 13:44:56.903392 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.606143 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsbbw" event={"ID":"0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6","Type":"ContainerDied","Data":"c74f7a8b7ab0e3f37ef2ae936c5f539034e20dd9a81db59b3381c8571e8dbe74"} Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.606468 4778 scope.go:117] "RemoveContainer" containerID="53b4b05091c277f65d760aedbdd673f748699fdc3c86d1184ac2f57494d75369" Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.606242 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsbbw" Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.629506 4778 scope.go:117] "RemoveContainer" containerID="e4ef6232820924fc9bfa9e22782695ff3623174a4e886e71c18dbcccc1dd4b88" Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.641558 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsbbw"] Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.652676 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsbbw"] Mar 12 13:44:57 crc kubenswrapper[4778]: I0312 13:44:57.676414 4778 scope.go:117] "RemoveContainer" containerID="b80555cb1dcab02445567c3e276216686d7719c0ca0475be3608a7a9c0f26718" Mar 12 13:44:58 crc kubenswrapper[4778]: I0312 13:44:58.269662 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" path="/var/lib/kubelet/pods/0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6/volumes" Mar 12 13:44:58 crc kubenswrapper[4778]: I0312 13:44:58.557798 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:44:58 crc kubenswrapper[4778]: I0312 13:44:58.558255 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:44:58 crc kubenswrapper[4778]: I0312 13:44:58.558399 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:44:58 crc kubenswrapper[4778]: I0312 13:44:58.559401 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92d3dad2e98d7139cb748a76fe93295a7064a4a757626bc932a272018a133968"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:44:58 crc kubenswrapper[4778]: I0312 13:44:58.559584 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://92d3dad2e98d7139cb748a76fe93295a7064a4a757626bc932a272018a133968" gracePeriod=600 Mar 12 13:44:59 crc kubenswrapper[4778]: I0312 13:44:59.628337 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="92d3dad2e98d7139cb748a76fe93295a7064a4a757626bc932a272018a133968" exitCode=0 Mar 12 13:44:59 crc kubenswrapper[4778]: I0312 13:44:59.628433 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"92d3dad2e98d7139cb748a76fe93295a7064a4a757626bc932a272018a133968"} Mar 12 13:44:59 crc kubenswrapper[4778]: I0312 13:44:59.628855 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f"} Mar 12 13:44:59 crc kubenswrapper[4778]: I0312 13:44:59.628880 4778 scope.go:117] "RemoveContainer" containerID="fbdf0765f9c2ff5952a8a2a2b43d61ef771ac404cabeb86051f9ffe5a9fd882e" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.150378 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh"] Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151138 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="extract-content" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151153 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="extract-content" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151166 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="extract-content" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151173 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="extract-content" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151203 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151211 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151223 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151228 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151247 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151253 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151263 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="extract-utilities" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151269 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="extract-utilities" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151278 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="extract-utilities" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151292 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="extract-utilities" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151301 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="extract-content" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151308 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="extract-content" Mar 12 13:45:00 crc kubenswrapper[4778]: E0312 13:45:00.151326 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="extract-utilities" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151332 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="extract-utilities" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151505 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d361e0-0808-41e0-a659-d9977bce86de" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151520 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f91915-3841-4662-88e4-82a22df0b131" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.151533 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f9b6c4c-ccc9-42ac-9d52-64690d25a4d6" containerName="registry-server" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.152142 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.155328 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.159310 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.165500 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh"] Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.168812 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76005d52-2d02-4a1e-89dd-c050a66fe667-secret-volume\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.168886 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqdwc\" (UniqueName: \"kubernetes.io/projected/76005d52-2d02-4a1e-89dd-c050a66fe667-kube-api-access-rqdwc\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.168937 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76005d52-2d02-4a1e-89dd-c050a66fe667-config-volume\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.271095 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76005d52-2d02-4a1e-89dd-c050a66fe667-secret-volume\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.271241 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqdwc\" (UniqueName: \"kubernetes.io/projected/76005d52-2d02-4a1e-89dd-c050a66fe667-kube-api-access-rqdwc\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.271324 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76005d52-2d02-4a1e-89dd-c050a66fe667-config-volume\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.272237 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76005d52-2d02-4a1e-89dd-c050a66fe667-config-volume\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.281863 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76005d52-2d02-4a1e-89dd-c050a66fe667-secret-volume\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.293960 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqdwc\" (UniqueName: \"kubernetes.io/projected/76005d52-2d02-4a1e-89dd-c050a66fe667-kube-api-access-rqdwc\") pod \"collect-profiles-29555385-qwzwh\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:00 crc kubenswrapper[4778]: I0312 13:45:00.473472 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:01 crc kubenswrapper[4778]: I0312 13:45:01.074843 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh"] Mar 12 13:45:01 crc kubenswrapper[4778]: E0312 13:45:01.531587 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76005d52_2d02_4a1e_89dd_c050a66fe667.slice/crio-conmon-501e76905e9d2ff1f1e87040184d63ca0f219b530ef232d95f1fa4250e5ab145.scope\": RecentStats: unable to find data in memory cache]" Mar 12 13:45:01 crc kubenswrapper[4778]: I0312 13:45:01.654841 4778 generic.go:334] "Generic (PLEG): container finished" podID="76005d52-2d02-4a1e-89dd-c050a66fe667" containerID="501e76905e9d2ff1f1e87040184d63ca0f219b530ef232d95f1fa4250e5ab145" exitCode=0 Mar 12 13:45:01 crc kubenswrapper[4778]: I0312 13:45:01.654910 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" event={"ID":"76005d52-2d02-4a1e-89dd-c050a66fe667","Type":"ContainerDied","Data":"501e76905e9d2ff1f1e87040184d63ca0f219b530ef232d95f1fa4250e5ab145"} Mar 12 13:45:01 crc kubenswrapper[4778]: I0312 13:45:01.654941 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" event={"ID":"76005d52-2d02-4a1e-89dd-c050a66fe667","Type":"ContainerStarted","Data":"924e154bac07decf2e0d0ef1ee072b8e965f02c42cadb7e7817f33bcd904899a"} Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.018614 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.130745 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76005d52-2d02-4a1e-89dd-c050a66fe667-config-volume\") pod \"76005d52-2d02-4a1e-89dd-c050a66fe667\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.130872 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76005d52-2d02-4a1e-89dd-c050a66fe667-secret-volume\") pod \"76005d52-2d02-4a1e-89dd-c050a66fe667\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.131001 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqdwc\" (UniqueName: \"kubernetes.io/projected/76005d52-2d02-4a1e-89dd-c050a66fe667-kube-api-access-rqdwc\") pod \"76005d52-2d02-4a1e-89dd-c050a66fe667\" (UID: \"76005d52-2d02-4a1e-89dd-c050a66fe667\") " Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.131754 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76005d52-2d02-4a1e-89dd-c050a66fe667-config-volume" (OuterVolumeSpecName: "config-volume") pod "76005d52-2d02-4a1e-89dd-c050a66fe667" (UID: "76005d52-2d02-4a1e-89dd-c050a66fe667"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.132304 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76005d52-2d02-4a1e-89dd-c050a66fe667-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.139042 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76005d52-2d02-4a1e-89dd-c050a66fe667-kube-api-access-rqdwc" (OuterVolumeSpecName: "kube-api-access-rqdwc") pod "76005d52-2d02-4a1e-89dd-c050a66fe667" (UID: "76005d52-2d02-4a1e-89dd-c050a66fe667"). InnerVolumeSpecName "kube-api-access-rqdwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.139237 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76005d52-2d02-4a1e-89dd-c050a66fe667-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "76005d52-2d02-4a1e-89dd-c050a66fe667" (UID: "76005d52-2d02-4a1e-89dd-c050a66fe667"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.234651 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76005d52-2d02-4a1e-89dd-c050a66fe667-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.234703 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqdwc\" (UniqueName: \"kubernetes.io/projected/76005d52-2d02-4a1e-89dd-c050a66fe667-kube-api-access-rqdwc\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.675898 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.675828 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh" event={"ID":"76005d52-2d02-4a1e-89dd-c050a66fe667","Type":"ContainerDied","Data":"924e154bac07decf2e0d0ef1ee072b8e965f02c42cadb7e7817f33bcd904899a"} Mar 12 13:45:03 crc kubenswrapper[4778]: I0312 13:45:03.676341 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="924e154bac07decf2e0d0ef1ee072b8e965f02c42cadb7e7817f33bcd904899a" Mar 12 13:45:04 crc kubenswrapper[4778]: I0312 13:45:04.091309 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm"] Mar 12 13:45:04 crc kubenswrapper[4778]: I0312 13:45:04.099415 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555340-7tvjm"] Mar 12 13:45:04 crc kubenswrapper[4778]: I0312 13:45:04.267135 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a240fd7b-5854-4548-a847-e5590111964b" path="/var/lib/kubelet/pods/a240fd7b-5854-4548-a847-e5590111964b/volumes" Mar 12 13:45:18 crc kubenswrapper[4778]: I0312 13:45:18.653693 4778 scope.go:117] "RemoveContainer" containerID="638395848d77320f6f4d74ca6334a62beda4c18b92408c089881a124597a1418" Mar 12 13:45:18 crc kubenswrapper[4778]: I0312 13:45:18.721630 4778 scope.go:117] "RemoveContainer" containerID="a7c208f5185dc692f0ec8df98f6bb0b7b464e0a056d454057e864768b033e299" Mar 12 13:45:18 crc kubenswrapper[4778]: I0312 13:45:18.787430 4778 scope.go:117] "RemoveContainer" containerID="e3b15e2b52f4e1dd648d1cbcdd4c757ead8e48ae1ed5c998744e64dfa8993e67" Mar 12 13:45:18 crc kubenswrapper[4778]: I0312 13:45:18.874659 4778 scope.go:117] "RemoveContainer" containerID="6e9a4135f2199a3918c9a565e1055b2ed771be6904f7c3aed074108524b55a58" Mar 12 13:45:30 crc kubenswrapper[4778]: I0312 13:45:30.935212 4778 generic.go:334] "Generic (PLEG): container finished" podID="5c5541f3-fb44-476b-91c2-b07dffe50894" containerID="f8a3ca31df85cf9da8faf63b19eca49715e39783019bbe431b8fd2f9f7f72be3" exitCode=0 Mar 12 13:45:30 crc kubenswrapper[4778]: I0312 13:45:30.935549 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" event={"ID":"5c5541f3-fb44-476b-91c2-b07dffe50894","Type":"ContainerDied","Data":"f8a3ca31df85cf9da8faf63b19eca49715e39783019bbe431b8fd2f9f7f72be3"} Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.308602 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.418968 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-inventory\") pod \"5c5541f3-fb44-476b-91c2-b07dffe50894\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.419095 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-ssh-key-openstack-edpm-ipam\") pod \"5c5541f3-fb44-476b-91c2-b07dffe50894\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.419243 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lntxg\" (UniqueName: \"kubernetes.io/projected/5c5541f3-fb44-476b-91c2-b07dffe50894-kube-api-access-lntxg\") pod \"5c5541f3-fb44-476b-91c2-b07dffe50894\" (UID: \"5c5541f3-fb44-476b-91c2-b07dffe50894\") " Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.429381 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c5541f3-fb44-476b-91c2-b07dffe50894-kube-api-access-lntxg" (OuterVolumeSpecName: "kube-api-access-lntxg") pod "5c5541f3-fb44-476b-91c2-b07dffe50894" (UID: "5c5541f3-fb44-476b-91c2-b07dffe50894"). InnerVolumeSpecName "kube-api-access-lntxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.447502 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-inventory" (OuterVolumeSpecName: "inventory") pod "5c5541f3-fb44-476b-91c2-b07dffe50894" (UID: "5c5541f3-fb44-476b-91c2-b07dffe50894"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.448459 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5c5541f3-fb44-476b-91c2-b07dffe50894" (UID: "5c5541f3-fb44-476b-91c2-b07dffe50894"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.521549 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.521590 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5c5541f3-fb44-476b-91c2-b07dffe50894-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.521603 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lntxg\" (UniqueName: \"kubernetes.io/projected/5c5541f3-fb44-476b-91c2-b07dffe50894-kube-api-access-lntxg\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.956413 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" event={"ID":"5c5541f3-fb44-476b-91c2-b07dffe50894","Type":"ContainerDied","Data":"2a03fc2efe0354ad242b25364cbd2eba2b14ebdbcdcea201fdb32f5a13b5d430"} Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.956468 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a03fc2efe0354ad242b25364cbd2eba2b14ebdbcdcea201fdb32f5a13b5d430" Mar 12 13:45:32 crc kubenswrapper[4778]: I0312 13:45:32.956521 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4szjl" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.073722 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr"] Mar 12 13:45:33 crc kubenswrapper[4778]: E0312 13:45:33.074308 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c5541f3-fb44-476b-91c2-b07dffe50894" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.074332 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c5541f3-fb44-476b-91c2-b07dffe50894" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 12 13:45:33 crc kubenswrapper[4778]: E0312 13:45:33.074350 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76005d52-2d02-4a1e-89dd-c050a66fe667" containerName="collect-profiles" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.074359 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="76005d52-2d02-4a1e-89dd-c050a66fe667" containerName="collect-profiles" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.074582 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c5541f3-fb44-476b-91c2-b07dffe50894" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.074603 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="76005d52-2d02-4a1e-89dd-c050a66fe667" containerName="collect-profiles" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.075385 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.084742 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr"] Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.088549 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.088645 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.088855 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.088557 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.132015 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.132362 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr45c\" (UniqueName: \"kubernetes.io/projected/41583476-38cd-4c0d-a05a-96ddc5b330ca-kube-api-access-gr45c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.132571 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.233941 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.234017 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr45c\" (UniqueName: \"kubernetes.io/projected/41583476-38cd-4c0d-a05a-96ddc5b330ca-kube-api-access-gr45c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.234082 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.240235 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.241852 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.255974 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr45c\" (UniqueName: \"kubernetes.io/projected/41583476-38cd-4c0d-a05a-96ddc5b330ca-kube-api-access-gr45c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9glvr\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.391555 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.897112 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr"] Mar 12 13:45:33 crc kubenswrapper[4778]: I0312 13:45:33.964240 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" event={"ID":"41583476-38cd-4c0d-a05a-96ddc5b330ca","Type":"ContainerStarted","Data":"359850f324feb41d74844acf8dd8291a291310bc46c176a33198262535f0d3da"} Mar 12 13:45:35 crc kubenswrapper[4778]: I0312 13:45:35.981855 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" event={"ID":"41583476-38cd-4c0d-a05a-96ddc5b330ca","Type":"ContainerStarted","Data":"d3aa4c129131d7684c185c26239b44c165d321201a6c5511e96be838b7de91e5"} Mar 12 13:45:36 crc kubenswrapper[4778]: I0312 13:45:36.008298 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" podStartSLOduration=1.287300717 podStartE2EDuration="3.008278015s" podCreationTimestamp="2026-03-12 13:45:33 +0000 UTC" firstStartedPulling="2026-03-12 13:45:33.903453796 +0000 UTC m=+2152.352149202" lastFinishedPulling="2026-03-12 13:45:35.624431104 +0000 UTC m=+2154.073126500" observedRunningTime="2026-03-12 13:45:35.996522931 +0000 UTC m=+2154.445218347" watchObservedRunningTime="2026-03-12 13:45:36.008278015 +0000 UTC m=+2154.456973411" Mar 12 13:45:39 crc kubenswrapper[4778]: I0312 13:45:39.059284 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9xw6b"] Mar 12 13:45:39 crc kubenswrapper[4778]: I0312 13:45:39.070495 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9xw6b"] Mar 12 13:45:40 crc kubenswrapper[4778]: I0312 13:45:40.266442 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3" path="/var/lib/kubelet/pods/eaa0985c-3171-40c5-8e5c-ab82a9fa6fc3/volumes" Mar 12 13:45:45 crc kubenswrapper[4778]: I0312 13:45:45.456460 4778 generic.go:334] "Generic (PLEG): container finished" podID="41583476-38cd-4c0d-a05a-96ddc5b330ca" containerID="d3aa4c129131d7684c185c26239b44c165d321201a6c5511e96be838b7de91e5" exitCode=0 Mar 12 13:45:45 crc kubenswrapper[4778]: I0312 13:45:45.456594 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" event={"ID":"41583476-38cd-4c0d-a05a-96ddc5b330ca","Type":"ContainerDied","Data":"d3aa4c129131d7684c185c26239b44c165d321201a6c5511e96be838b7de91e5"} Mar 12 13:45:46 crc kubenswrapper[4778]: I0312 13:45:46.903209 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:46 crc kubenswrapper[4778]: I0312 13:45:46.991212 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-ssh-key-openstack-edpm-ipam\") pod \"41583476-38cd-4c0d-a05a-96ddc5b330ca\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " Mar 12 13:45:46 crc kubenswrapper[4778]: I0312 13:45:46.991781 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-inventory\") pod \"41583476-38cd-4c0d-a05a-96ddc5b330ca\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " Mar 12 13:45:46 crc kubenswrapper[4778]: I0312 13:45:46.991905 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr45c\" (UniqueName: \"kubernetes.io/projected/41583476-38cd-4c0d-a05a-96ddc5b330ca-kube-api-access-gr45c\") pod \"41583476-38cd-4c0d-a05a-96ddc5b330ca\" (UID: \"41583476-38cd-4c0d-a05a-96ddc5b330ca\") " Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:46.999872 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41583476-38cd-4c0d-a05a-96ddc5b330ca-kube-api-access-gr45c" (OuterVolumeSpecName: "kube-api-access-gr45c") pod "41583476-38cd-4c0d-a05a-96ddc5b330ca" (UID: "41583476-38cd-4c0d-a05a-96ddc5b330ca"). InnerVolumeSpecName "kube-api-access-gr45c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.029814 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-inventory" (OuterVolumeSpecName: "inventory") pod "41583476-38cd-4c0d-a05a-96ddc5b330ca" (UID: "41583476-38cd-4c0d-a05a-96ddc5b330ca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.033378 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "41583476-38cd-4c0d-a05a-96ddc5b330ca" (UID: "41583476-38cd-4c0d-a05a-96ddc5b330ca"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.096433 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.096553 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41583476-38cd-4c0d-a05a-96ddc5b330ca-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.096571 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr45c\" (UniqueName: \"kubernetes.io/projected/41583476-38cd-4c0d-a05a-96ddc5b330ca-kube-api-access-gr45c\") on node \"crc\" DevicePath \"\"" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.475617 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" event={"ID":"41583476-38cd-4c0d-a05a-96ddc5b330ca","Type":"ContainerDied","Data":"359850f324feb41d74844acf8dd8291a291310bc46c176a33198262535f0d3da"} Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.475669 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="359850f324feb41d74844acf8dd8291a291310bc46c176a33198262535f0d3da" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.475732 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9glvr" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.573049 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n"] Mar 12 13:45:47 crc kubenswrapper[4778]: E0312 13:45:47.574156 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41583476-38cd-4c0d-a05a-96ddc5b330ca" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.574317 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="41583476-38cd-4c0d-a05a-96ddc5b330ca" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.574626 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="41583476-38cd-4c0d-a05a-96ddc5b330ca" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.575563 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.578441 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.578834 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.580363 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.580798 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.587392 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n"] Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.707493 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.707604 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.707837 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zh8c\" (UniqueName: \"kubernetes.io/projected/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-kube-api-access-5zh8c\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.809632 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zh8c\" (UniqueName: \"kubernetes.io/projected/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-kube-api-access-5zh8c\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.809732 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.809778 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.813555 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.814168 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.830642 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zh8c\" (UniqueName: \"kubernetes.io/projected/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-kube-api-access-5zh8c\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g252n\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:47 crc kubenswrapper[4778]: I0312 13:45:47.907154 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:45:48 crc kubenswrapper[4778]: I0312 13:45:48.474436 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n"] Mar 12 13:45:48 crc kubenswrapper[4778]: W0312 13:45:48.481210 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29f8609b_4a3b_42ba_9450_a2b633bb4c2c.slice/crio-73a22b7b256ee030de16c8af304a9e966b8c86513de32e0f1f1efbf8d69bce4e WatchSource:0}: Error finding container 73a22b7b256ee030de16c8af304a9e966b8c86513de32e0f1f1efbf8d69bce4e: Status 404 returned error can't find the container with id 73a22b7b256ee030de16c8af304a9e966b8c86513de32e0f1f1efbf8d69bce4e Mar 12 13:45:49 crc kubenswrapper[4778]: I0312 13:45:49.491766 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" event={"ID":"29f8609b-4a3b-42ba-9450-a2b633bb4c2c","Type":"ContainerStarted","Data":"6d041d9afe88fdec32af7da1355c78c05ba3b4da3265df6c00f4f7416ed27298"} Mar 12 13:45:49 crc kubenswrapper[4778]: I0312 13:45:49.492341 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" event={"ID":"29f8609b-4a3b-42ba-9450-a2b633bb4c2c","Type":"ContainerStarted","Data":"73a22b7b256ee030de16c8af304a9e966b8c86513de32e0f1f1efbf8d69bce4e"} Mar 12 13:45:50 crc kubenswrapper[4778]: I0312 13:45:50.524987 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" podStartSLOduration=2.928433387 podStartE2EDuration="3.524969979s" podCreationTimestamp="2026-03-12 13:45:47 +0000 UTC" firstStartedPulling="2026-03-12 13:45:48.483989564 +0000 UTC m=+2166.932684960" lastFinishedPulling="2026-03-12 13:45:49.080526156 +0000 UTC m=+2167.529221552" observedRunningTime="2026-03-12 13:45:50.51584531 +0000 UTC m=+2168.964540696" watchObservedRunningTime="2026-03-12 13:45:50.524969979 +0000 UTC m=+2168.973665375" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.146135 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555386-vjswk"] Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.149410 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.153959 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.154284 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.155340 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.155980 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555386-vjswk"] Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.245759 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8gz7\" (UniqueName: \"kubernetes.io/projected/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6-kube-api-access-c8gz7\") pod \"auto-csr-approver-29555386-vjswk\" (UID: \"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6\") " pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.349676 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8gz7\" (UniqueName: \"kubernetes.io/projected/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6-kube-api-access-c8gz7\") pod \"auto-csr-approver-29555386-vjswk\" (UID: \"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6\") " pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.371021 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8gz7\" (UniqueName: \"kubernetes.io/projected/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6-kube-api-access-c8gz7\") pod \"auto-csr-approver-29555386-vjswk\" (UID: \"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6\") " pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.487096 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:00 crc kubenswrapper[4778]: I0312 13:46:00.937861 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555386-vjswk"] Mar 12 13:46:00 crc kubenswrapper[4778]: W0312 13:46:00.944009 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55c85e9_4cb7_4ac4_bc3d_c37217b4abf6.slice/crio-615af91b8c037a610414459977df85a0b58a194b765f6201a453a8ec1331a26e WatchSource:0}: Error finding container 615af91b8c037a610414459977df85a0b58a194b765f6201a453a8ec1331a26e: Status 404 returned error can't find the container with id 615af91b8c037a610414459977df85a0b58a194b765f6201a453a8ec1331a26e Mar 12 13:46:01 crc kubenswrapper[4778]: I0312 13:46:01.591845 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555386-vjswk" event={"ID":"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6","Type":"ContainerStarted","Data":"615af91b8c037a610414459977df85a0b58a194b765f6201a453a8ec1331a26e"} Mar 12 13:46:04 crc kubenswrapper[4778]: I0312 13:46:04.935978 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l8q2d"] Mar 12 13:46:04 crc kubenswrapper[4778]: I0312 13:46:04.939061 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:04 crc kubenswrapper[4778]: I0312 13:46:04.946353 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8q2d"] Mar 12 13:46:04 crc kubenswrapper[4778]: I0312 13:46:04.983278 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-utilities\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:04 crc kubenswrapper[4778]: I0312 13:46:04.983328 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-catalog-content\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:04 crc kubenswrapper[4778]: I0312 13:46:04.983617 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfslw\" (UniqueName: \"kubernetes.io/projected/af355b7f-362d-4f00-96fb-07f77590df88-kube-api-access-gfslw\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.021506 4778 generic.go:334] "Generic (PLEG): container finished" podID="f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6" containerID="2e424e585231dad361491fa12a9a787f83d6973879b6b45159764198bbcf5877" exitCode=0 Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.021564 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555386-vjswk" event={"ID":"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6","Type":"ContainerDied","Data":"2e424e585231dad361491fa12a9a787f83d6973879b6b45159764198bbcf5877"} Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.085693 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-utilities\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.085779 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-catalog-content\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.085856 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfslw\" (UniqueName: \"kubernetes.io/projected/af355b7f-362d-4f00-96fb-07f77590df88-kube-api-access-gfslw\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.086430 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-utilities\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.086722 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-catalog-content\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.111700 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfslw\" (UniqueName: \"kubernetes.io/projected/af355b7f-362d-4f00-96fb-07f77590df88-kube-api-access-gfslw\") pod \"redhat-operators-l8q2d\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.255038 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:05 crc kubenswrapper[4778]: I0312 13:46:05.718033 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8q2d"] Mar 12 13:46:06 crc kubenswrapper[4778]: I0312 13:46:06.030563 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerStarted","Data":"d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f"} Mar 12 13:46:06 crc kubenswrapper[4778]: I0312 13:46:06.030611 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerStarted","Data":"32b7530aa6acd71d36db41ee270d32aa3aa01b7f38534e4287b54333a34c61e6"} Mar 12 13:46:06 crc kubenswrapper[4778]: I0312 13:46:06.428240 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:06 crc kubenswrapper[4778]: I0312 13:46:06.616258 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8gz7\" (UniqueName: \"kubernetes.io/projected/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6-kube-api-access-c8gz7\") pod \"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6\" (UID: \"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6\") " Mar 12 13:46:06 crc kubenswrapper[4778]: I0312 13:46:06.624264 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6-kube-api-access-c8gz7" (OuterVolumeSpecName: "kube-api-access-c8gz7") pod "f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6" (UID: "f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6"). InnerVolumeSpecName "kube-api-access-c8gz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:46:06 crc kubenswrapper[4778]: I0312 13:46:06.719518 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8gz7\" (UniqueName: \"kubernetes.io/projected/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6-kube-api-access-c8gz7\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.042062 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555386-vjswk" Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.042056 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555386-vjswk" event={"ID":"f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6","Type":"ContainerDied","Data":"615af91b8c037a610414459977df85a0b58a194b765f6201a453a8ec1331a26e"} Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.042235 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="615af91b8c037a610414459977df85a0b58a194b765f6201a453a8ec1331a26e" Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.059979 4778 generic.go:334] "Generic (PLEG): container finished" podID="af355b7f-362d-4f00-96fb-07f77590df88" containerID="d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f" exitCode=0 Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.060043 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerDied","Data":"d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f"} Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.515801 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555380-n8mtp"] Mar 12 13:46:07 crc kubenswrapper[4778]: I0312 13:46:07.525952 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555380-n8mtp"] Mar 12 13:46:08 crc kubenswrapper[4778]: I0312 13:46:08.265803 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f54cc7-08e2-42c1-883d-316f1dac7621" path="/var/lib/kubelet/pods/69f54cc7-08e2-42c1-883d-316f1dac7621/volumes" Mar 12 13:46:09 crc kubenswrapper[4778]: I0312 13:46:09.096564 4778 generic.go:334] "Generic (PLEG): container finished" podID="af355b7f-362d-4f00-96fb-07f77590df88" containerID="d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1" exitCode=0 Mar 12 13:46:09 crc kubenswrapper[4778]: I0312 13:46:09.096621 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerDied","Data":"d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1"} Mar 12 13:46:11 crc kubenswrapper[4778]: I0312 13:46:11.121264 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerStarted","Data":"7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6"} Mar 12 13:46:11 crc kubenswrapper[4778]: I0312 13:46:11.142309 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l8q2d" podStartSLOduration=4.071254563 podStartE2EDuration="7.142292192s" podCreationTimestamp="2026-03-12 13:46:04 +0000 UTC" firstStartedPulling="2026-03-12 13:46:07.06236947 +0000 UTC m=+2185.511064886" lastFinishedPulling="2026-03-12 13:46:10.133407119 +0000 UTC m=+2188.582102515" observedRunningTime="2026-03-12 13:46:11.141674794 +0000 UTC m=+2189.590370200" watchObservedRunningTime="2026-03-12 13:46:11.142292192 +0000 UTC m=+2189.590987588" Mar 12 13:46:15 crc kubenswrapper[4778]: I0312 13:46:15.255966 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:15 crc kubenswrapper[4778]: I0312 13:46:15.256677 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:16 crc kubenswrapper[4778]: I0312 13:46:16.306113 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l8q2d" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="registry-server" probeResult="failure" output=< Mar 12 13:46:16 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:46:16 crc kubenswrapper[4778]: > Mar 12 13:46:19 crc kubenswrapper[4778]: I0312 13:46:19.020594 4778 scope.go:117] "RemoveContainer" containerID="a3547232ddc46df5ded5cc24fff2ec3e7c8bb4fb4c52277d66e27c319ec41995" Mar 12 13:46:19 crc kubenswrapper[4778]: I0312 13:46:19.068513 4778 scope.go:117] "RemoveContainer" containerID="625dea5df6820f4416903072a858eb0ac8d225248a71973001f9856768eaad43" Mar 12 13:46:25 crc kubenswrapper[4778]: I0312 13:46:25.309924 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:25 crc kubenswrapper[4778]: I0312 13:46:25.367215 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:25 crc kubenswrapper[4778]: I0312 13:46:25.555069 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l8q2d"] Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.250358 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l8q2d" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="registry-server" containerID="cri-o://7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6" gracePeriod=2 Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.770054 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.856958 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfslw\" (UniqueName: \"kubernetes.io/projected/af355b7f-362d-4f00-96fb-07f77590df88-kube-api-access-gfslw\") pod \"af355b7f-362d-4f00-96fb-07f77590df88\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.857143 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-catalog-content\") pod \"af355b7f-362d-4f00-96fb-07f77590df88\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.857258 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-utilities\") pod \"af355b7f-362d-4f00-96fb-07f77590df88\" (UID: \"af355b7f-362d-4f00-96fb-07f77590df88\") " Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.858329 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-utilities" (OuterVolumeSpecName: "utilities") pod "af355b7f-362d-4f00-96fb-07f77590df88" (UID: "af355b7f-362d-4f00-96fb-07f77590df88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.873948 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af355b7f-362d-4f00-96fb-07f77590df88-kube-api-access-gfslw" (OuterVolumeSpecName: "kube-api-access-gfslw") pod "af355b7f-362d-4f00-96fb-07f77590df88" (UID: "af355b7f-362d-4f00-96fb-07f77590df88"). InnerVolumeSpecName "kube-api-access-gfslw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.959444 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.959483 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfslw\" (UniqueName: \"kubernetes.io/projected/af355b7f-362d-4f00-96fb-07f77590df88-kube-api-access-gfslw\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:27 crc kubenswrapper[4778]: I0312 13:46:27.993617 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af355b7f-362d-4f00-96fb-07f77590df88" (UID: "af355b7f-362d-4f00-96fb-07f77590df88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.061949 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af355b7f-362d-4f00-96fb-07f77590df88-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.267661 4778 generic.go:334] "Generic (PLEG): container finished" podID="af355b7f-362d-4f00-96fb-07f77590df88" containerID="7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6" exitCode=0 Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.267940 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8q2d" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.274946 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerDied","Data":"7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6"} Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.275009 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8q2d" event={"ID":"af355b7f-362d-4f00-96fb-07f77590df88","Type":"ContainerDied","Data":"32b7530aa6acd71d36db41ee270d32aa3aa01b7f38534e4287b54333a34c61e6"} Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.275032 4778 scope.go:117] "RemoveContainer" containerID="7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.299259 4778 scope.go:117] "RemoveContainer" containerID="d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.317778 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l8q2d"] Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.326620 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l8q2d"] Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.339320 4778 scope.go:117] "RemoveContainer" containerID="d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.376478 4778 scope.go:117] "RemoveContainer" containerID="7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6" Mar 12 13:46:28 crc kubenswrapper[4778]: E0312 13:46:28.377127 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6\": container with ID starting with 7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6 not found: ID does not exist" containerID="7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.377195 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6"} err="failed to get container status \"7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6\": rpc error: code = NotFound desc = could not find container \"7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6\": container with ID starting with 7eb94c616cc863bfc286172d8cd2e64a1dfd8876e97f1d5ec12fdbf103ee91e6 not found: ID does not exist" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.377222 4778 scope.go:117] "RemoveContainer" containerID="d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1" Mar 12 13:46:28 crc kubenswrapper[4778]: E0312 13:46:28.377805 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1\": container with ID starting with d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1 not found: ID does not exist" containerID="d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.377894 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1"} err="failed to get container status \"d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1\": rpc error: code = NotFound desc = could not find container \"d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1\": container with ID starting with d99e4622ec7401bc478f28e049daf1fe07e501d51d99bc26762c1751e79641f1 not found: ID does not exist" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.377945 4778 scope.go:117] "RemoveContainer" containerID="d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f" Mar 12 13:46:28 crc kubenswrapper[4778]: E0312 13:46:28.378554 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f\": container with ID starting with d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f not found: ID does not exist" containerID="d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f" Mar 12 13:46:28 crc kubenswrapper[4778]: I0312 13:46:28.378592 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f"} err="failed to get container status \"d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f\": rpc error: code = NotFound desc = could not find container \"d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f\": container with ID starting with d774e6cca00651044cc9a3c3d37f2a22c2a44d3dee64303589589539c636c16f not found: ID does not exist" Mar 12 13:46:30 crc kubenswrapper[4778]: I0312 13:46:30.264628 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af355b7f-362d-4f00-96fb-07f77590df88" path="/var/lib/kubelet/pods/af355b7f-362d-4f00-96fb-07f77590df88/volumes" Mar 12 13:46:33 crc kubenswrapper[4778]: I0312 13:46:33.323104 4778 generic.go:334] "Generic (PLEG): container finished" podID="29f8609b-4a3b-42ba-9450-a2b633bb4c2c" containerID="6d041d9afe88fdec32af7da1355c78c05ba3b4da3265df6c00f4f7416ed27298" exitCode=0 Mar 12 13:46:33 crc kubenswrapper[4778]: I0312 13:46:33.323151 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" event={"ID":"29f8609b-4a3b-42ba-9450-a2b633bb4c2c","Type":"ContainerDied","Data":"6d041d9afe88fdec32af7da1355c78c05ba3b4da3265df6c00f4f7416ed27298"} Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.832104 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.920824 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-ssh-key-openstack-edpm-ipam\") pod \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.920986 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zh8c\" (UniqueName: \"kubernetes.io/projected/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-kube-api-access-5zh8c\") pod \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.921015 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-inventory\") pod \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\" (UID: \"29f8609b-4a3b-42ba-9450-a2b633bb4c2c\") " Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.928887 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-kube-api-access-5zh8c" (OuterVolumeSpecName: "kube-api-access-5zh8c") pod "29f8609b-4a3b-42ba-9450-a2b633bb4c2c" (UID: "29f8609b-4a3b-42ba-9450-a2b633bb4c2c"). InnerVolumeSpecName "kube-api-access-5zh8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.950957 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "29f8609b-4a3b-42ba-9450-a2b633bb4c2c" (UID: "29f8609b-4a3b-42ba-9450-a2b633bb4c2c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:46:34 crc kubenswrapper[4778]: I0312 13:46:34.957501 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-inventory" (OuterVolumeSpecName: "inventory") pod "29f8609b-4a3b-42ba-9450-a2b633bb4c2c" (UID: "29f8609b-4a3b-42ba-9450-a2b633bb4c2c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.024267 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.024303 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zh8c\" (UniqueName: \"kubernetes.io/projected/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-kube-api-access-5zh8c\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.024315 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29f8609b-4a3b-42ba-9450-a2b633bb4c2c-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.339848 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" event={"ID":"29f8609b-4a3b-42ba-9450-a2b633bb4c2c","Type":"ContainerDied","Data":"73a22b7b256ee030de16c8af304a9e966b8c86513de32e0f1f1efbf8d69bce4e"} Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.339902 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73a22b7b256ee030de16c8af304a9e966b8c86513de32e0f1f1efbf8d69bce4e" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.339994 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g252n" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.425167 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6"] Mar 12 13:46:35 crc kubenswrapper[4778]: E0312 13:46:35.426588 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6" containerName="oc" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.426677 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6" containerName="oc" Mar 12 13:46:35 crc kubenswrapper[4778]: E0312 13:46:35.426750 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="registry-server" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.426815 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="registry-server" Mar 12 13:46:35 crc kubenswrapper[4778]: E0312 13:46:35.426884 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29f8609b-4a3b-42ba-9450-a2b633bb4c2c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.426942 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="29f8609b-4a3b-42ba-9450-a2b633bb4c2c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:46:35 crc kubenswrapper[4778]: E0312 13:46:35.427013 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="extract-utilities" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.427071 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="extract-utilities" Mar 12 13:46:35 crc kubenswrapper[4778]: E0312 13:46:35.427144 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="extract-content" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.427223 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="extract-content" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.427465 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="af355b7f-362d-4f00-96fb-07f77590df88" containerName="registry-server" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.427539 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6" containerName="oc" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.427613 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="29f8609b-4a3b-42ba-9450-a2b633bb4c2c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.428290 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.436303 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.436593 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.437989 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.439448 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6"] Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.445964 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.539956 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.540044 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.540089 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m74f9\" (UniqueName: \"kubernetes.io/projected/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-kube-api-access-m74f9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.642401 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.642804 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.642879 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m74f9\" (UniqueName: \"kubernetes.io/projected/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-kube-api-access-m74f9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.648672 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.649073 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.657992 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m74f9\" (UniqueName: \"kubernetes.io/projected/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-kube-api-access-m74f9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:35 crc kubenswrapper[4778]: I0312 13:46:35.749701 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:46:36 crc kubenswrapper[4778]: I0312 13:46:36.281441 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6"] Mar 12 13:46:36 crc kubenswrapper[4778]: I0312 13:46:36.350300 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" event={"ID":"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8","Type":"ContainerStarted","Data":"cf09e698453bb9cd997120f8269dfad3a83af17a3258f80df0e30f8d276af66c"} Mar 12 13:46:37 crc kubenswrapper[4778]: I0312 13:46:37.358900 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" event={"ID":"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8","Type":"ContainerStarted","Data":"e50c9beda337639076d5173ca63e16ac54e3a0b42a8b327eda05f2300a28b3fe"} Mar 12 13:46:37 crc kubenswrapper[4778]: I0312 13:46:37.382537 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" podStartSLOduration=1.8789279159999999 podStartE2EDuration="2.382517518s" podCreationTimestamp="2026-03-12 13:46:35 +0000 UTC" firstStartedPulling="2026-03-12 13:46:36.287096448 +0000 UTC m=+2214.735791834" lastFinishedPulling="2026-03-12 13:46:36.79068604 +0000 UTC m=+2215.239381436" observedRunningTime="2026-03-12 13:46:37.376072354 +0000 UTC m=+2215.824767750" watchObservedRunningTime="2026-03-12 13:46:37.382517518 +0000 UTC m=+2215.831212914" Mar 12 13:46:58 crc kubenswrapper[4778]: I0312 13:46:58.557972 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:46:58 crc kubenswrapper[4778]: I0312 13:46:58.559328 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:47:25 crc kubenswrapper[4778]: I0312 13:47:25.572827 4778 generic.go:334] "Generic (PLEG): container finished" podID="36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" containerID="e50c9beda337639076d5173ca63e16ac54e3a0b42a8b327eda05f2300a28b3fe" exitCode=0 Mar 12 13:47:25 crc kubenswrapper[4778]: I0312 13:47:25.572917 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" event={"ID":"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8","Type":"ContainerDied","Data":"e50c9beda337639076d5173ca63e16ac54e3a0b42a8b327eda05f2300a28b3fe"} Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.008325 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.141400 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m74f9\" (UniqueName: \"kubernetes.io/projected/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-kube-api-access-m74f9\") pod \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.141600 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-inventory\") pod \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.141754 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-ssh-key-openstack-edpm-ipam\") pod \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\" (UID: \"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8\") " Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.150774 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-kube-api-access-m74f9" (OuterVolumeSpecName: "kube-api-access-m74f9") pod "36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" (UID: "36bb4acd-fab3-4998-a8cd-a6ebcc800fc8"). InnerVolumeSpecName "kube-api-access-m74f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.172025 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" (UID: "36bb4acd-fab3-4998-a8cd-a6ebcc800fc8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.172505 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-inventory" (OuterVolumeSpecName: "inventory") pod "36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" (UID: "36bb4acd-fab3-4998-a8cd-a6ebcc800fc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.243600 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m74f9\" (UniqueName: \"kubernetes.io/projected/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-kube-api-access-m74f9\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.243641 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.243654 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36bb4acd-fab3-4998-a8cd-a6ebcc800fc8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.593418 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" event={"ID":"36bb4acd-fab3-4998-a8cd-a6ebcc800fc8","Type":"ContainerDied","Data":"cf09e698453bb9cd997120f8269dfad3a83af17a3258f80df0e30f8d276af66c"} Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.593512 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf09e698453bb9cd997120f8269dfad3a83af17a3258f80df0e30f8d276af66c" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.593439 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.694886 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8mmjm"] Mar 12 13:47:27 crc kubenswrapper[4778]: E0312 13:47:27.695355 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.695371 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.695539 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="36bb4acd-fab3-4998-a8cd-a6ebcc800fc8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.696160 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.699147 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.699336 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.699458 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.699686 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.714236 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8mmjm"] Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.856328 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.856725 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.857122 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxkkp\" (UniqueName: \"kubernetes.io/projected/c993b33e-6c36-4524-864a-65da461a8e0c-kube-api-access-mxkkp\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.959083 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxkkp\" (UniqueName: \"kubernetes.io/projected/c993b33e-6c36-4524-864a-65da461a8e0c-kube-api-access-mxkkp\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.959217 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.959291 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.965489 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.966953 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:27 crc kubenswrapper[4778]: I0312 13:47:27.979211 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxkkp\" (UniqueName: \"kubernetes.io/projected/c993b33e-6c36-4524-864a-65da461a8e0c-kube-api-access-mxkkp\") pod \"ssh-known-hosts-edpm-deployment-8mmjm\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:28 crc kubenswrapper[4778]: I0312 13:47:28.015397 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:28 crc kubenswrapper[4778]: I0312 13:47:28.526104 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8mmjm"] Mar 12 13:47:28 crc kubenswrapper[4778]: I0312 13:47:28.557618 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:47:28 crc kubenswrapper[4778]: I0312 13:47:28.557695 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:47:28 crc kubenswrapper[4778]: I0312 13:47:28.601891 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" event={"ID":"c993b33e-6c36-4524-864a-65da461a8e0c","Type":"ContainerStarted","Data":"40a927d39fc02c94f89eec5982f76353c83245af8433d34ae91ec6311691714a"} Mar 12 13:47:29 crc kubenswrapper[4778]: I0312 13:47:29.614076 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" event={"ID":"c993b33e-6c36-4524-864a-65da461a8e0c","Type":"ContainerStarted","Data":"4db62d3ee6fc5f306e1af4017f76ced458e8adef04cfa2db62bf3d06afae11ed"} Mar 12 13:47:29 crc kubenswrapper[4778]: I0312 13:47:29.636458 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" podStartSLOduration=2.052265158 podStartE2EDuration="2.636437029s" podCreationTimestamp="2026-03-12 13:47:27 +0000 UTC" firstStartedPulling="2026-03-12 13:47:28.533944187 +0000 UTC m=+2266.982639593" lastFinishedPulling="2026-03-12 13:47:29.118116068 +0000 UTC m=+2267.566811464" observedRunningTime="2026-03-12 13:47:29.628175914 +0000 UTC m=+2268.076871310" watchObservedRunningTime="2026-03-12 13:47:29.636437029 +0000 UTC m=+2268.085132415" Mar 12 13:47:36 crc kubenswrapper[4778]: I0312 13:47:36.676581 4778 generic.go:334] "Generic (PLEG): container finished" podID="c993b33e-6c36-4524-864a-65da461a8e0c" containerID="4db62d3ee6fc5f306e1af4017f76ced458e8adef04cfa2db62bf3d06afae11ed" exitCode=0 Mar 12 13:47:36 crc kubenswrapper[4778]: I0312 13:47:36.676708 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" event={"ID":"c993b33e-6c36-4524-864a-65da461a8e0c","Type":"ContainerDied","Data":"4db62d3ee6fc5f306e1af4017f76ced458e8adef04cfa2db62bf3d06afae11ed"} Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.146072 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.272145 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxkkp\" (UniqueName: \"kubernetes.io/projected/c993b33e-6c36-4524-864a-65da461a8e0c-kube-api-access-mxkkp\") pod \"c993b33e-6c36-4524-864a-65da461a8e0c\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.272313 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-ssh-key-openstack-edpm-ipam\") pod \"c993b33e-6c36-4524-864a-65da461a8e0c\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.272358 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-inventory-0\") pod \"c993b33e-6c36-4524-864a-65da461a8e0c\" (UID: \"c993b33e-6c36-4524-864a-65da461a8e0c\") " Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.277269 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c993b33e-6c36-4524-864a-65da461a8e0c-kube-api-access-mxkkp" (OuterVolumeSpecName: "kube-api-access-mxkkp") pod "c993b33e-6c36-4524-864a-65da461a8e0c" (UID: "c993b33e-6c36-4524-864a-65da461a8e0c"). InnerVolumeSpecName "kube-api-access-mxkkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.299026 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "c993b33e-6c36-4524-864a-65da461a8e0c" (UID: "c993b33e-6c36-4524-864a-65da461a8e0c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.304111 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c993b33e-6c36-4524-864a-65da461a8e0c" (UID: "c993b33e-6c36-4524-864a-65da461a8e0c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.375682 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.375734 4778 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c993b33e-6c36-4524-864a-65da461a8e0c-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.375749 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxkkp\" (UniqueName: \"kubernetes.io/projected/c993b33e-6c36-4524-864a-65da461a8e0c-kube-api-access-mxkkp\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.694950 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" event={"ID":"c993b33e-6c36-4524-864a-65da461a8e0c","Type":"ContainerDied","Data":"40a927d39fc02c94f89eec5982f76353c83245af8433d34ae91ec6311691714a"} Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.694992 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40a927d39fc02c94f89eec5982f76353c83245af8433d34ae91ec6311691714a" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.695047 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8mmjm" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.781414 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t"] Mar 12 13:47:38 crc kubenswrapper[4778]: E0312 13:47:38.782145 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c993b33e-6c36-4524-864a-65da461a8e0c" containerName="ssh-known-hosts-edpm-deployment" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.782169 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c993b33e-6c36-4524-864a-65da461a8e0c" containerName="ssh-known-hosts-edpm-deployment" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.782441 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c993b33e-6c36-4524-864a-65da461a8e0c" containerName="ssh-known-hosts-edpm-deployment" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.783421 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.785747 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.786006 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.788373 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.788563 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.790617 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t"] Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.987649 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.987963 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:38 crc kubenswrapper[4778]: I0312 13:47:38.988007 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrmsd\" (UniqueName: \"kubernetes.io/projected/b0bb06df-44bb-4939-9492-a6ad3d6b5368-kube-api-access-lrmsd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.090032 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.090089 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.090123 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrmsd\" (UniqueName: \"kubernetes.io/projected/b0bb06df-44bb-4939-9492-a6ad3d6b5368-kube-api-access-lrmsd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.094091 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.094486 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.110230 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrmsd\" (UniqueName: \"kubernetes.io/projected/b0bb06df-44bb-4939-9492-a6ad3d6b5368-kube-api-access-lrmsd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gt58t\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.401812 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:39 crc kubenswrapper[4778]: I0312 13:47:39.917054 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t"] Mar 12 13:47:40 crc kubenswrapper[4778]: I0312 13:47:40.718787 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" event={"ID":"b0bb06df-44bb-4939-9492-a6ad3d6b5368","Type":"ContainerStarted","Data":"4406f329a8888ee9cb9a4e349401d009362172ceafbfcf53efeaa42f370ffb2e"} Mar 12 13:47:41 crc kubenswrapper[4778]: I0312 13:47:41.729018 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" event={"ID":"b0bb06df-44bb-4939-9492-a6ad3d6b5368","Type":"ContainerStarted","Data":"2e5a4191c8140186b288301cee49d0ac609d54636f3611c2d402c382f0805fb8"} Mar 12 13:47:41 crc kubenswrapper[4778]: I0312 13:47:41.745766 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" podStartSLOduration=3.157034611 podStartE2EDuration="3.745745831s" podCreationTimestamp="2026-03-12 13:47:38 +0000 UTC" firstStartedPulling="2026-03-12 13:47:39.925609377 +0000 UTC m=+2278.374304773" lastFinishedPulling="2026-03-12 13:47:40.514320597 +0000 UTC m=+2278.963015993" observedRunningTime="2026-03-12 13:47:41.744072493 +0000 UTC m=+2280.192767899" watchObservedRunningTime="2026-03-12 13:47:41.745745831 +0000 UTC m=+2280.194441227" Mar 12 13:47:48 crc kubenswrapper[4778]: I0312 13:47:48.791824 4778 generic.go:334] "Generic (PLEG): container finished" podID="b0bb06df-44bb-4939-9492-a6ad3d6b5368" containerID="2e5a4191c8140186b288301cee49d0ac609d54636f3611c2d402c382f0805fb8" exitCode=0 Mar 12 13:47:48 crc kubenswrapper[4778]: I0312 13:47:48.791916 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" event={"ID":"b0bb06df-44bb-4939-9492-a6ad3d6b5368","Type":"ContainerDied","Data":"2e5a4191c8140186b288301cee49d0ac609d54636f3611c2d402c382f0805fb8"} Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.228268 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.424080 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-inventory\") pod \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.424588 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-ssh-key-openstack-edpm-ipam\") pod \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.424628 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrmsd\" (UniqueName: \"kubernetes.io/projected/b0bb06df-44bb-4939-9492-a6ad3d6b5368-kube-api-access-lrmsd\") pod \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\" (UID: \"b0bb06df-44bb-4939-9492-a6ad3d6b5368\") " Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.431214 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0bb06df-44bb-4939-9492-a6ad3d6b5368-kube-api-access-lrmsd" (OuterVolumeSpecName: "kube-api-access-lrmsd") pod "b0bb06df-44bb-4939-9492-a6ad3d6b5368" (UID: "b0bb06df-44bb-4939-9492-a6ad3d6b5368"). InnerVolumeSpecName "kube-api-access-lrmsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.454455 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-inventory" (OuterVolumeSpecName: "inventory") pod "b0bb06df-44bb-4939-9492-a6ad3d6b5368" (UID: "b0bb06df-44bb-4939-9492-a6ad3d6b5368"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.459619 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b0bb06df-44bb-4939-9492-a6ad3d6b5368" (UID: "b0bb06df-44bb-4939-9492-a6ad3d6b5368"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.529432 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.529468 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrmsd\" (UniqueName: \"kubernetes.io/projected/b0bb06df-44bb-4939-9492-a6ad3d6b5368-kube-api-access-lrmsd\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.529477 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0bb06df-44bb-4939-9492-a6ad3d6b5368-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.817777 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" event={"ID":"b0bb06df-44bb-4939-9492-a6ad3d6b5368","Type":"ContainerDied","Data":"4406f329a8888ee9cb9a4e349401d009362172ceafbfcf53efeaa42f370ffb2e"} Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.817844 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4406f329a8888ee9cb9a4e349401d009362172ceafbfcf53efeaa42f370ffb2e" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.817855 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gt58t" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.914076 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc"] Mar 12 13:47:50 crc kubenswrapper[4778]: E0312 13:47:50.914778 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0bb06df-44bb-4939-9492-a6ad3d6b5368" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.914808 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0bb06df-44bb-4939-9492-a6ad3d6b5368" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.915051 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0bb06df-44bb-4939-9492-a6ad3d6b5368" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.915904 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.918279 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.919455 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.919611 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.919649 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:47:50 crc kubenswrapper[4778]: I0312 13:47:50.925619 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc"] Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.039046 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.039444 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqrq9\" (UniqueName: \"kubernetes.io/projected/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-kube-api-access-rqrq9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.039481 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.140778 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.140840 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqrq9\" (UniqueName: \"kubernetes.io/projected/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-kube-api-access-rqrq9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.140864 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.145440 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.146614 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.159385 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqrq9\" (UniqueName: \"kubernetes.io/projected/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-kube-api-access-rqrq9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.237446 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.738839 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc"] Mar 12 13:47:51 crc kubenswrapper[4778]: I0312 13:47:51.828136 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" event={"ID":"43a3ffe4-8b64-4e26-b63a-5254a986e4a4","Type":"ContainerStarted","Data":"feca20715b50a556b5694815d65a26e8fe73431d82cde071b3fb63cd519e73aa"} Mar 12 13:47:52 crc kubenswrapper[4778]: I0312 13:47:52.839418 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" event={"ID":"43a3ffe4-8b64-4e26-b63a-5254a986e4a4","Type":"ContainerStarted","Data":"7e12694d9161e0f60ee30d919973a8d39ab3d9a3f7092129e17f862f3f4116b9"} Mar 12 13:47:52 crc kubenswrapper[4778]: I0312 13:47:52.858671 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" podStartSLOduration=2.3163973909999998 podStartE2EDuration="2.858651852s" podCreationTimestamp="2026-03-12 13:47:50 +0000 UTC" firstStartedPulling="2026-03-12 13:47:51.739015674 +0000 UTC m=+2290.187711070" lastFinishedPulling="2026-03-12 13:47:52.281270135 +0000 UTC m=+2290.729965531" observedRunningTime="2026-03-12 13:47:52.855152573 +0000 UTC m=+2291.303847979" watchObservedRunningTime="2026-03-12 13:47:52.858651852 +0000 UTC m=+2291.307347248" Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.558585 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.559192 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.559274 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.560293 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.560370 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" gracePeriod=600 Mar 12 13:47:58 crc kubenswrapper[4778]: E0312 13:47:58.683345 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.891032 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" exitCode=0 Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.891100 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f"} Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.891467 4778 scope.go:117] "RemoveContainer" containerID="92d3dad2e98d7139cb748a76fe93295a7064a4a757626bc932a272018a133968" Mar 12 13:47:58 crc kubenswrapper[4778]: I0312 13:47:58.892132 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:47:58 crc kubenswrapper[4778]: E0312 13:47:58.892493 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.132379 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555388-5mnjh"] Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.134788 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.137293 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.139702 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.140146 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.155314 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555388-5mnjh"] Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.207990 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqhkb\" (UniqueName: \"kubernetes.io/projected/110071e6-5231-434c-af16-87b68a3d0c8f-kube-api-access-pqhkb\") pod \"auto-csr-approver-29555388-5mnjh\" (UID: \"110071e6-5231-434c-af16-87b68a3d0c8f\") " pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.310285 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqhkb\" (UniqueName: \"kubernetes.io/projected/110071e6-5231-434c-af16-87b68a3d0c8f-kube-api-access-pqhkb\") pod \"auto-csr-approver-29555388-5mnjh\" (UID: \"110071e6-5231-434c-af16-87b68a3d0c8f\") " pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.331375 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqhkb\" (UniqueName: \"kubernetes.io/projected/110071e6-5231-434c-af16-87b68a3d0c8f-kube-api-access-pqhkb\") pod \"auto-csr-approver-29555388-5mnjh\" (UID: \"110071e6-5231-434c-af16-87b68a3d0c8f\") " pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.457982 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.891998 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555388-5mnjh"] Mar 12 13:48:00 crc kubenswrapper[4778]: I0312 13:48:00.912559 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" event={"ID":"110071e6-5231-434c-af16-87b68a3d0c8f","Type":"ContainerStarted","Data":"f9bbce7ec9ea3b75de4f987d83a468acfd1a280c283c8c43f14d3ba69def0c98"} Mar 12 13:48:01 crc kubenswrapper[4778]: I0312 13:48:01.921713 4778 generic.go:334] "Generic (PLEG): container finished" podID="43a3ffe4-8b64-4e26-b63a-5254a986e4a4" containerID="7e12694d9161e0f60ee30d919973a8d39ab3d9a3f7092129e17f862f3f4116b9" exitCode=0 Mar 12 13:48:01 crc kubenswrapper[4778]: I0312 13:48:01.921805 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" event={"ID":"43a3ffe4-8b64-4e26-b63a-5254a986e4a4","Type":"ContainerDied","Data":"7e12694d9161e0f60ee30d919973a8d39ab3d9a3f7092129e17f862f3f4116b9"} Mar 12 13:48:02 crc kubenswrapper[4778]: I0312 13:48:02.932108 4778 generic.go:334] "Generic (PLEG): container finished" podID="110071e6-5231-434c-af16-87b68a3d0c8f" containerID="2b6df075041c6c1583e329716aacbd0c53d1a64cadc9905cc6ddb1e1bd9b676d" exitCode=0 Mar 12 13:48:02 crc kubenswrapper[4778]: I0312 13:48:02.932260 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" event={"ID":"110071e6-5231-434c-af16-87b68a3d0c8f","Type":"ContainerDied","Data":"2b6df075041c6c1583e329716aacbd0c53d1a64cadc9905cc6ddb1e1bd9b676d"} Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.364155 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.483785 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-inventory\") pod \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.483881 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqrq9\" (UniqueName: \"kubernetes.io/projected/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-kube-api-access-rqrq9\") pod \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.483971 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-ssh-key-openstack-edpm-ipam\") pod \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\" (UID: \"43a3ffe4-8b64-4e26-b63a-5254a986e4a4\") " Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.492001 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-kube-api-access-rqrq9" (OuterVolumeSpecName: "kube-api-access-rqrq9") pod "43a3ffe4-8b64-4e26-b63a-5254a986e4a4" (UID: "43a3ffe4-8b64-4e26-b63a-5254a986e4a4"). InnerVolumeSpecName "kube-api-access-rqrq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.515252 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "43a3ffe4-8b64-4e26-b63a-5254a986e4a4" (UID: "43a3ffe4-8b64-4e26-b63a-5254a986e4a4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.526717 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-inventory" (OuterVolumeSpecName: "inventory") pod "43a3ffe4-8b64-4e26-b63a-5254a986e4a4" (UID: "43a3ffe4-8b64-4e26-b63a-5254a986e4a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.586564 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.586592 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqrq9\" (UniqueName: \"kubernetes.io/projected/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-kube-api-access-rqrq9\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.586605 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/43a3ffe4-8b64-4e26-b63a-5254a986e4a4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.946827 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" event={"ID":"43a3ffe4-8b64-4e26-b63a-5254a986e4a4","Type":"ContainerDied","Data":"feca20715b50a556b5694815d65a26e8fe73431d82cde071b3fb63cd519e73aa"} Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.946879 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feca20715b50a556b5694815d65a26e8fe73431d82cde071b3fb63cd519e73aa" Mar 12 13:48:03 crc kubenswrapper[4778]: I0312 13:48:03.946842 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.075697 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx"] Mar 12 13:48:04 crc kubenswrapper[4778]: E0312 13:48:04.076298 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a3ffe4-8b64-4e26-b63a-5254a986e4a4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.076323 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a3ffe4-8b64-4e26-b63a-5254a986e4a4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.076607 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a3ffe4-8b64-4e26-b63a-5254a986e4a4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.077489 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.083484 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.083718 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.084099 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.084215 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-custom-default-certs-0" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.084577 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.084708 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.084874 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.084986 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.095413 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx"] Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.201762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.201939 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202097 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202197 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202233 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202317 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202645 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202679 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202825 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202868 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202955 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-neutron-metadata-custom-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.202981 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.203075 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8pb\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-kube-api-access-bl8pb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.203103 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.305102 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306013 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306058 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-neutron-metadata-custom-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306078 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306118 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8pb\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-kube-api-access-bl8pb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306136 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306235 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306270 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306314 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306342 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306360 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306389 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306451 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.306470 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.311655 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.311836 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.311976 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.311283 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.312755 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.312898 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.313387 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-neutron-metadata-custom-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.313858 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.314565 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.315264 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.315912 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.316127 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.325917 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.329153 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8pb\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-kube-api-access-bl8pb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bngcx\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.407697 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.421943 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.511696 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqhkb\" (UniqueName: \"kubernetes.io/projected/110071e6-5231-434c-af16-87b68a3d0c8f-kube-api-access-pqhkb\") pod \"110071e6-5231-434c-af16-87b68a3d0c8f\" (UID: \"110071e6-5231-434c-af16-87b68a3d0c8f\") " Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.516054 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/110071e6-5231-434c-af16-87b68a3d0c8f-kube-api-access-pqhkb" (OuterVolumeSpecName: "kube-api-access-pqhkb") pod "110071e6-5231-434c-af16-87b68a3d0c8f" (UID: "110071e6-5231-434c-af16-87b68a3d0c8f"). InnerVolumeSpecName "kube-api-access-pqhkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.614099 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqhkb\" (UniqueName: \"kubernetes.io/projected/110071e6-5231-434c-af16-87b68a3d0c8f-kube-api-access-pqhkb\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.933541 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx"] Mar 12 13:48:04 crc kubenswrapper[4778]: W0312 13:48:04.934855 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf69e6cfe_f7c2_4127_b4df_710725c52227.slice/crio-201137bf7718cef671d660ab8f7e18f0a7ed8f9c84776ebbb382aa1672d08d37 WatchSource:0}: Error finding container 201137bf7718cef671d660ab8f7e18f0a7ed8f9c84776ebbb382aa1672d08d37: Status 404 returned error can't find the container with id 201137bf7718cef671d660ab8f7e18f0a7ed8f9c84776ebbb382aa1672d08d37 Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.965468 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" event={"ID":"110071e6-5231-434c-af16-87b68a3d0c8f","Type":"ContainerDied","Data":"f9bbce7ec9ea3b75de4f987d83a468acfd1a280c283c8c43f14d3ba69def0c98"} Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.965518 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9bbce7ec9ea3b75de4f987d83a468acfd1a280c283c8c43f14d3ba69def0c98" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.965515 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555388-5mnjh" Mar 12 13:48:04 crc kubenswrapper[4778]: I0312 13:48:04.966469 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" event={"ID":"f69e6cfe-f7c2-4127-b4df-710725c52227","Type":"ContainerStarted","Data":"201137bf7718cef671d660ab8f7e18f0a7ed8f9c84776ebbb382aa1672d08d37"} Mar 12 13:48:05 crc kubenswrapper[4778]: I0312 13:48:05.550697 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555382-zbkfk"] Mar 12 13:48:05 crc kubenswrapper[4778]: I0312 13:48:05.559804 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555382-zbkfk"] Mar 12 13:48:05 crc kubenswrapper[4778]: I0312 13:48:05.989067 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" event={"ID":"f69e6cfe-f7c2-4127-b4df-710725c52227","Type":"ContainerStarted","Data":"1b6058478ca276578c95fa23d5a23fc397088e6b19d0e6a8d4aa362015a60ea0"} Mar 12 13:48:06 crc kubenswrapper[4778]: I0312 13:48:06.025127 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" podStartSLOduration=1.399275345 podStartE2EDuration="2.025098249s" podCreationTimestamp="2026-03-12 13:48:04 +0000 UTC" firstStartedPulling="2026-03-12 13:48:04.938693224 +0000 UTC m=+2303.387388620" lastFinishedPulling="2026-03-12 13:48:05.564516128 +0000 UTC m=+2304.013211524" observedRunningTime="2026-03-12 13:48:06.018322137 +0000 UTC m=+2304.467017553" watchObservedRunningTime="2026-03-12 13:48:06.025098249 +0000 UTC m=+2304.473793645" Mar 12 13:48:06 crc kubenswrapper[4778]: I0312 13:48:06.277140 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="832c789c-468c-400b-8d55-3072443e85ec" path="/var/lib/kubelet/pods/832c789c-468c-400b-8d55-3072443e85ec/volumes" Mar 12 13:48:13 crc kubenswrapper[4778]: I0312 13:48:13.254827 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:48:13 crc kubenswrapper[4778]: E0312 13:48:13.255524 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:48:19 crc kubenswrapper[4778]: I0312 13:48:19.198826 4778 scope.go:117] "RemoveContainer" containerID="7785d6a0c6670e984508e3f9d5cc59f211b972f130207a3fed5c63411c140ddc" Mar 12 13:48:26 crc kubenswrapper[4778]: I0312 13:48:26.255021 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:48:26 crc kubenswrapper[4778]: E0312 13:48:26.255594 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:48:37 crc kubenswrapper[4778]: I0312 13:48:37.254576 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:48:37 crc kubenswrapper[4778]: E0312 13:48:37.255433 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:48:42 crc kubenswrapper[4778]: I0312 13:48:42.594536 4778 generic.go:334] "Generic (PLEG): container finished" podID="f69e6cfe-f7c2-4127-b4df-710725c52227" containerID="1b6058478ca276578c95fa23d5a23fc397088e6b19d0e6a8d4aa362015a60ea0" exitCode=0 Mar 12 13:48:42 crc kubenswrapper[4778]: I0312 13:48:42.594582 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" event={"ID":"f69e6cfe-f7c2-4127-b4df-710725c52227","Type":"ContainerDied","Data":"1b6058478ca276578c95fa23d5a23fc397088e6b19d0e6a8d4aa362015a60ea0"} Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.046057 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.085710 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ssh-key-openstack-edpm-ipam\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.085985 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-repo-setup-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086091 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-inventory\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086204 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-bootstrap-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086290 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086367 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086444 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-telemetry-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086534 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-nova-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086628 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl8pb\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-kube-api-access-bl8pb\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086741 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-neutron-metadata-custom-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086815 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-libvirt-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.086891 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.094061 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.097302 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.097509 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-neutron-metadata-custom-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-custom-default-certs-0") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-custom-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.097552 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-kube-api-access-bl8pb" (OuterVolumeSpecName: "kube-api-access-bl8pb") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "kube-api-access-bl8pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.098088 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.098165 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.099526 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.100511 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.100607 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-neutron-metadata-custom-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-custom-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "neutron-metadata-custom-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.101338 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.128112 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.128593 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-inventory" (OuterVolumeSpecName: "inventory") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188006 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188390 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ovn-combined-ca-bundle\") pod \"f69e6cfe-f7c2-4127-b4df-710725c52227\" (UID: \"f69e6cfe-f7c2-4127-b4df-710725c52227\") " Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188776 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188802 4778 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188817 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188832 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-neutron-metadata-custom-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188847 4778 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188861 4778 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188871 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl8pb\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-kube-api-access-bl8pb\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188882 4778 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188893 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188908 4778 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-neutron-metadata-custom-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188919 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.188929 4778 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.191511 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.191832 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f69e6cfe-f7c2-4127-b4df-710725c52227" (UID: "f69e6cfe-f7c2-4127-b4df-710725c52227"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.291991 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f69e6cfe-f7c2-4127-b4df-710725c52227-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.292041 4778 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6cfe-f7c2-4127-b4df-710725c52227-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.802343 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" event={"ID":"f69e6cfe-f7c2-4127-b4df-710725c52227","Type":"ContainerDied","Data":"201137bf7718cef671d660ab8f7e18f0a7ed8f9c84776ebbb382aa1672d08d37"} Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.802387 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="201137bf7718cef671d660ab8f7e18f0a7ed8f9c84776ebbb382aa1672d08d37" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.802478 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bngcx" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.978988 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq"] Mar 12 13:48:44 crc kubenswrapper[4778]: E0312 13:48:44.979441 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="110071e6-5231-434c-af16-87b68a3d0c8f" containerName="oc" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.979464 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="110071e6-5231-434c-af16-87b68a3d0c8f" containerName="oc" Mar 12 13:48:44 crc kubenswrapper[4778]: E0312 13:48:44.979487 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69e6cfe-f7c2-4127-b4df-710725c52227" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.979499 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69e6cfe-f7c2-4127-b4df-710725c52227" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.979757 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="110071e6-5231-434c-af16-87b68a3d0c8f" containerName="oc" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.979798 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69e6cfe-f7c2-4127-b4df-710725c52227" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.980662 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.983649 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.984005 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.984499 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.984604 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.985197 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:48:44 crc kubenswrapper[4778]: I0312 13:48:44.995301 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq"] Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.088436 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.088685 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wbdc\" (UniqueName: \"kubernetes.io/projected/3c0a2200-506d-4ac3-b08c-9b3156c9e573-kube-api-access-9wbdc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.088840 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.088905 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.088940 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.190641 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.190739 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wbdc\" (UniqueName: \"kubernetes.io/projected/3c0a2200-506d-4ac3-b08c-9b3156c9e573-kube-api-access-9wbdc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.190815 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.190846 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.190875 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.191881 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.195306 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.195548 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.210233 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.211502 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wbdc\" (UniqueName: \"kubernetes.io/projected/3c0a2200-506d-4ac3-b08c-9b3156c9e573-kube-api-access-9wbdc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9lbdq\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:45 crc kubenswrapper[4778]: I0312 13:48:45.304590 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:48:46 crc kubenswrapper[4778]: I0312 13:48:46.108618 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq"] Mar 12 13:48:47 crc kubenswrapper[4778]: I0312 13:48:47.063411 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" event={"ID":"3c0a2200-506d-4ac3-b08c-9b3156c9e573","Type":"ContainerStarted","Data":"9284931a0ee9d2beed15173491cfa2f871463185d1221d834da82de0b7e7f86f"} Mar 12 13:48:48 crc kubenswrapper[4778]: I0312 13:48:48.255212 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:48:48 crc kubenswrapper[4778]: E0312 13:48:48.255840 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:48:48 crc kubenswrapper[4778]: I0312 13:48:48.269927 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" event={"ID":"3c0a2200-506d-4ac3-b08c-9b3156c9e573","Type":"ContainerStarted","Data":"034471fb3d1e6422e07aa4976640757d8f19ab32e4556431d2f747552958f007"} Mar 12 13:48:48 crc kubenswrapper[4778]: I0312 13:48:48.291348 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" podStartSLOduration=3.422887011 podStartE2EDuration="4.291323284s" podCreationTimestamp="2026-03-12 13:48:44 +0000 UTC" firstStartedPulling="2026-03-12 13:48:46.114847401 +0000 UTC m=+2344.563542797" lastFinishedPulling="2026-03-12 13:48:46.983283674 +0000 UTC m=+2345.431979070" observedRunningTime="2026-03-12 13:48:48.284217322 +0000 UTC m=+2346.732912718" watchObservedRunningTime="2026-03-12 13:48:48.291323284 +0000 UTC m=+2346.740018680" Mar 12 13:49:01 crc kubenswrapper[4778]: I0312 13:49:01.254083 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:49:01 crc kubenswrapper[4778]: E0312 13:49:01.255047 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:49:12 crc kubenswrapper[4778]: I0312 13:49:12.253916 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:49:12 crc kubenswrapper[4778]: E0312 13:49:12.254839 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:49:25 crc kubenswrapper[4778]: I0312 13:49:25.254445 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:49:25 crc kubenswrapper[4778]: E0312 13:49:25.255417 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:49:40 crc kubenswrapper[4778]: I0312 13:49:40.254075 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:49:40 crc kubenswrapper[4778]: E0312 13:49:40.254906 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:49:54 crc kubenswrapper[4778]: I0312 13:49:54.254281 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:49:54 crc kubenswrapper[4778]: E0312 13:49:54.254967 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:49:54 crc kubenswrapper[4778]: I0312 13:49:54.270083 4778 generic.go:334] "Generic (PLEG): container finished" podID="3c0a2200-506d-4ac3-b08c-9b3156c9e573" containerID="034471fb3d1e6422e07aa4976640757d8f19ab32e4556431d2f747552958f007" exitCode=0 Mar 12 13:49:54 crc kubenswrapper[4778]: I0312 13:49:54.270123 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" event={"ID":"3c0a2200-506d-4ac3-b08c-9b3156c9e573","Type":"ContainerDied","Data":"034471fb3d1e6422e07aa4976640757d8f19ab32e4556431d2f747552958f007"} Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.725484 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.869738 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-inventory\") pod \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.869852 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovncontroller-config-0\") pod \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.869963 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovn-combined-ca-bundle\") pod \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.870023 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ssh-key-openstack-edpm-ipam\") pod \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.870043 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wbdc\" (UniqueName: \"kubernetes.io/projected/3c0a2200-506d-4ac3-b08c-9b3156c9e573-kube-api-access-9wbdc\") pod \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\" (UID: \"3c0a2200-506d-4ac3-b08c-9b3156c9e573\") " Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.876847 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0a2200-506d-4ac3-b08c-9b3156c9e573-kube-api-access-9wbdc" (OuterVolumeSpecName: "kube-api-access-9wbdc") pod "3c0a2200-506d-4ac3-b08c-9b3156c9e573" (UID: "3c0a2200-506d-4ac3-b08c-9b3156c9e573"). InnerVolumeSpecName "kube-api-access-9wbdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.881493 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3c0a2200-506d-4ac3-b08c-9b3156c9e573" (UID: "3c0a2200-506d-4ac3-b08c-9b3156c9e573"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.902843 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-inventory" (OuterVolumeSpecName: "inventory") pod "3c0a2200-506d-4ac3-b08c-9b3156c9e573" (UID: "3c0a2200-506d-4ac3-b08c-9b3156c9e573"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.904088 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3c0a2200-506d-4ac3-b08c-9b3156c9e573" (UID: "3c0a2200-506d-4ac3-b08c-9b3156c9e573"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.906031 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3c0a2200-506d-4ac3-b08c-9b3156c9e573" (UID: "3c0a2200-506d-4ac3-b08c-9b3156c9e573"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.972799 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.972990 4778 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.973067 4778 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.973127 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c0a2200-506d-4ac3-b08c-9b3156c9e573-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:49:55 crc kubenswrapper[4778]: I0312 13:49:55.973203 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wbdc\" (UniqueName: \"kubernetes.io/projected/3c0a2200-506d-4ac3-b08c-9b3156c9e573-kube-api-access-9wbdc\") on node \"crc\" DevicePath \"\"" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.308866 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" event={"ID":"3c0a2200-506d-4ac3-b08c-9b3156c9e573","Type":"ContainerDied","Data":"9284931a0ee9d2beed15173491cfa2f871463185d1221d834da82de0b7e7f86f"} Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.308912 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9284931a0ee9d2beed15173491cfa2f871463185d1221d834da82de0b7e7f86f" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.308914 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9lbdq" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.384736 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg"] Mar 12 13:49:56 crc kubenswrapper[4778]: E0312 13:49:56.385155 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0a2200-506d-4ac3-b08c-9b3156c9e573" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.385175 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0a2200-506d-4ac3-b08c-9b3156c9e573" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.385433 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0a2200-506d-4ac3-b08c-9b3156c9e573" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.386118 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.476494 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.476810 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.476957 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.477137 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-metadata-neutron-config" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.477298 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.477430 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.499504 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg"] Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.585860 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.586039 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8wvt\" (UniqueName: \"kubernetes.io/projected/5cc410de-5b42-44d1-8b29-37161475730e-kube-api-access-g8wvt\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.586348 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.586482 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-metadata-custom-combined-ca-bundle\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.586789 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-inventory\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.586913 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-nova-cell1-metadata-neutron-config-0\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.689233 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8wvt\" (UniqueName: \"kubernetes.io/projected/5cc410de-5b42-44d1-8b29-37161475730e-kube-api-access-g8wvt\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.689357 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.689396 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-metadata-custom-combined-ca-bundle\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.689551 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-inventory\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.689615 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-nova-cell1-metadata-neutron-config-0\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.689679 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.695097 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.695444 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-nova-cell1-metadata-neutron-config-0\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.695922 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-metadata-custom-combined-ca-bundle\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.698253 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.704781 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-inventory\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.709793 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8wvt\" (UniqueName: \"kubernetes.io/projected/5cc410de-5b42-44d1-8b29-37161475730e-kube-api-access-g8wvt\") pod \"neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:56 crc kubenswrapper[4778]: I0312 13:49:56.797816 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:49:57 crc kubenswrapper[4778]: I0312 13:49:57.372841 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg"] Mar 12 13:49:57 crc kubenswrapper[4778]: I0312 13:49:57.377311 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:49:58 crc kubenswrapper[4778]: I0312 13:49:58.330123 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" event={"ID":"5cc410de-5b42-44d1-8b29-37161475730e","Type":"ContainerStarted","Data":"8d307b414f96fb01f36bf7a3a773e62c07543e84217803c38e669173f219bc57"} Mar 12 13:49:58 crc kubenswrapper[4778]: I0312 13:49:58.330588 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" event={"ID":"5cc410de-5b42-44d1-8b29-37161475730e","Type":"ContainerStarted","Data":"df0c491452d065b345698565a278b9fb265300119cb4f52d617e573795f6237b"} Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.132740 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" podStartSLOduration=3.540319285 podStartE2EDuration="4.132715181s" podCreationTimestamp="2026-03-12 13:49:56 +0000 UTC" firstStartedPulling="2026-03-12 13:49:57.377045658 +0000 UTC m=+2415.825741054" lastFinishedPulling="2026-03-12 13:49:57.969441564 +0000 UTC m=+2416.418136950" observedRunningTime="2026-03-12 13:49:58.354427549 +0000 UTC m=+2416.803122945" watchObservedRunningTime="2026-03-12 13:50:00.132715181 +0000 UTC m=+2418.581410577" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.136310 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555390-dml9r"] Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.138404 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.141163 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.143534 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.152327 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.154136 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555390-dml9r"] Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.162010 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgll9\" (UniqueName: \"kubernetes.io/projected/8b911b49-7b0f-48ef-9626-cd43d308d596-kube-api-access-fgll9\") pod \"auto-csr-approver-29555390-dml9r\" (UID: \"8b911b49-7b0f-48ef-9626-cd43d308d596\") " pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.264406 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgll9\" (UniqueName: \"kubernetes.io/projected/8b911b49-7b0f-48ef-9626-cd43d308d596-kube-api-access-fgll9\") pod \"auto-csr-approver-29555390-dml9r\" (UID: \"8b911b49-7b0f-48ef-9626-cd43d308d596\") " pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.293687 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgll9\" (UniqueName: \"kubernetes.io/projected/8b911b49-7b0f-48ef-9626-cd43d308d596-kube-api-access-fgll9\") pod \"auto-csr-approver-29555390-dml9r\" (UID: \"8b911b49-7b0f-48ef-9626-cd43d308d596\") " pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.466900 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:00 crc kubenswrapper[4778]: W0312 13:50:00.926763 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b911b49_7b0f_48ef_9626_cd43d308d596.slice/crio-2e2a7da2c9fca297f720655456df247f407a38064448b8b331c8699bf971e8a8 WatchSource:0}: Error finding container 2e2a7da2c9fca297f720655456df247f407a38064448b8b331c8699bf971e8a8: Status 404 returned error can't find the container with id 2e2a7da2c9fca297f720655456df247f407a38064448b8b331c8699bf971e8a8 Mar 12 13:50:00 crc kubenswrapper[4778]: I0312 13:50:00.927034 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555390-dml9r"] Mar 12 13:50:01 crc kubenswrapper[4778]: I0312 13:50:01.354843 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555390-dml9r" event={"ID":"8b911b49-7b0f-48ef-9626-cd43d308d596","Type":"ContainerStarted","Data":"2e2a7da2c9fca297f720655456df247f407a38064448b8b331c8699bf971e8a8"} Mar 12 13:50:03 crc kubenswrapper[4778]: I0312 13:50:03.374257 4778 generic.go:334] "Generic (PLEG): container finished" podID="8b911b49-7b0f-48ef-9626-cd43d308d596" containerID="63aaec5f1f507e8a81d7498ca66c0663cdb9bde37e98025da0e464b4ce6c885e" exitCode=0 Mar 12 13:50:03 crc kubenswrapper[4778]: I0312 13:50:03.374303 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555390-dml9r" event={"ID":"8b911b49-7b0f-48ef-9626-cd43d308d596","Type":"ContainerDied","Data":"63aaec5f1f507e8a81d7498ca66c0663cdb9bde37e98025da0e464b4ce6c885e"} Mar 12 13:50:04 crc kubenswrapper[4778]: I0312 13:50:04.700444 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:04 crc kubenswrapper[4778]: I0312 13:50:04.846585 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgll9\" (UniqueName: \"kubernetes.io/projected/8b911b49-7b0f-48ef-9626-cd43d308d596-kube-api-access-fgll9\") pod \"8b911b49-7b0f-48ef-9626-cd43d308d596\" (UID: \"8b911b49-7b0f-48ef-9626-cd43d308d596\") " Mar 12 13:50:04 crc kubenswrapper[4778]: I0312 13:50:04.854434 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b911b49-7b0f-48ef-9626-cd43d308d596-kube-api-access-fgll9" (OuterVolumeSpecName: "kube-api-access-fgll9") pod "8b911b49-7b0f-48ef-9626-cd43d308d596" (UID: "8b911b49-7b0f-48ef-9626-cd43d308d596"). InnerVolumeSpecName "kube-api-access-fgll9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:50:04 crc kubenswrapper[4778]: I0312 13:50:04.949670 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgll9\" (UniqueName: \"kubernetes.io/projected/8b911b49-7b0f-48ef-9626-cd43d308d596-kube-api-access-fgll9\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:05 crc kubenswrapper[4778]: I0312 13:50:05.393504 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555390-dml9r" event={"ID":"8b911b49-7b0f-48ef-9626-cd43d308d596","Type":"ContainerDied","Data":"2e2a7da2c9fca297f720655456df247f407a38064448b8b331c8699bf971e8a8"} Mar 12 13:50:05 crc kubenswrapper[4778]: I0312 13:50:05.393551 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e2a7da2c9fca297f720655456df247f407a38064448b8b331c8699bf971e8a8" Mar 12 13:50:05 crc kubenswrapper[4778]: I0312 13:50:05.393585 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555390-dml9r" Mar 12 13:50:05 crc kubenswrapper[4778]: I0312 13:50:05.774574 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555384-znhr8"] Mar 12 13:50:05 crc kubenswrapper[4778]: I0312 13:50:05.784599 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555384-znhr8"] Mar 12 13:50:06 crc kubenswrapper[4778]: I0312 13:50:06.264643 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70dc8f5a-da90-4090-b630-a6a7bd438f64" path="/var/lib/kubelet/pods/70dc8f5a-da90-4090-b630-a6a7bd438f64/volumes" Mar 12 13:50:07 crc kubenswrapper[4778]: I0312 13:50:07.254263 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:50:07 crc kubenswrapper[4778]: E0312 13:50:07.254628 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:50:19 crc kubenswrapper[4778]: I0312 13:50:19.553426 4778 scope.go:117] "RemoveContainer" containerID="e97aad250ae3960e7483df5290e0221b9fbbbe6a75ec4afcb92fd5c46ee60b01" Mar 12 13:50:21 crc kubenswrapper[4778]: I0312 13:50:21.254553 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:50:21 crc kubenswrapper[4778]: E0312 13:50:21.255400 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:50:33 crc kubenswrapper[4778]: I0312 13:50:33.253749 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:50:33 crc kubenswrapper[4778]: E0312 13:50:33.254350 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:50:47 crc kubenswrapper[4778]: I0312 13:50:47.254054 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:50:47 crc kubenswrapper[4778]: E0312 13:50:47.254882 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:50:51 crc kubenswrapper[4778]: I0312 13:50:51.385801 4778 generic.go:334] "Generic (PLEG): container finished" podID="5cc410de-5b42-44d1-8b29-37161475730e" containerID="8d307b414f96fb01f36bf7a3a773e62c07543e84217803c38e669173f219bc57" exitCode=0 Mar 12 13:50:51 crc kubenswrapper[4778]: I0312 13:50:51.385897 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" event={"ID":"5cc410de-5b42-44d1-8b29-37161475730e","Type":"ContainerDied","Data":"8d307b414f96fb01f36bf7a3a773e62c07543e84217803c38e669173f219bc57"} Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.835639 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.948058 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-inventory\") pod \"5cc410de-5b42-44d1-8b29-37161475730e\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.948176 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-nova-cell1-metadata-neutron-config-0\") pod \"5cc410de-5b42-44d1-8b29-37161475730e\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.948226 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"5cc410de-5b42-44d1-8b29-37161475730e\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.948328 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-metadata-custom-combined-ca-bundle\") pod \"5cc410de-5b42-44d1-8b29-37161475730e\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.948372 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-ssh-key-openstack-edpm-ipam\") pod \"5cc410de-5b42-44d1-8b29-37161475730e\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.948406 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8wvt\" (UniqueName: \"kubernetes.io/projected/5cc410de-5b42-44d1-8b29-37161475730e-kube-api-access-g8wvt\") pod \"5cc410de-5b42-44d1-8b29-37161475730e\" (UID: \"5cc410de-5b42-44d1-8b29-37161475730e\") " Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.954065 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-metadata-custom-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-custom-combined-ca-bundle") pod "5cc410de-5b42-44d1-8b29-37161475730e" (UID: "5cc410de-5b42-44d1-8b29-37161475730e"). InnerVolumeSpecName "neutron-metadata-custom-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.954618 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc410de-5b42-44d1-8b29-37161475730e-kube-api-access-g8wvt" (OuterVolumeSpecName: "kube-api-access-g8wvt") pod "5cc410de-5b42-44d1-8b29-37161475730e" (UID: "5cc410de-5b42-44d1-8b29-37161475730e"). InnerVolumeSpecName "kube-api-access-g8wvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.975806 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "5cc410de-5b42-44d1-8b29-37161475730e" (UID: "5cc410de-5b42-44d1-8b29-37161475730e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.976144 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5cc410de-5b42-44d1-8b29-37161475730e" (UID: "5cc410de-5b42-44d1-8b29-37161475730e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.982428 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-inventory" (OuterVolumeSpecName: "inventory") pod "5cc410de-5b42-44d1-8b29-37161475730e" (UID: "5cc410de-5b42-44d1-8b29-37161475730e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:50:52 crc kubenswrapper[4778]: I0312 13:50:52.982868 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-nova-cell1-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-cell1-metadata-neutron-config-0") pod "5cc410de-5b42-44d1-8b29-37161475730e" (UID: "5cc410de-5b42-44d1-8b29-37161475730e"). InnerVolumeSpecName "nova-cell1-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.051419 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8wvt\" (UniqueName: \"kubernetes.io/projected/5cc410de-5b42-44d1-8b29-37161475730e-kube-api-access-g8wvt\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.051478 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.051497 4778 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-nova-cell1-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.051518 4778 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.051535 4778 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-custom-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-neutron-metadata-custom-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.051549 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5cc410de-5b42-44d1-8b29-37161475730e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.401024 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" event={"ID":"5cc410de-5b42-44d1-8b29-37161475730e","Type":"ContainerDied","Data":"df0c491452d065b345698565a278b9fb265300119cb4f52d617e573795f6237b"} Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.401070 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df0c491452d065b345698565a278b9fb265300119cb4f52d617e573795f6237b" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.401146 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.583800 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8"] Mar 12 13:50:53 crc kubenswrapper[4778]: E0312 13:50:53.584221 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc410de-5b42-44d1-8b29-37161475730e" containerName="neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.584237 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc410de-5b42-44d1-8b29-37161475730e" containerName="neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam" Mar 12 13:50:53 crc kubenswrapper[4778]: E0312 13:50:53.584270 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b911b49-7b0f-48ef-9626-cd43d308d596" containerName="oc" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.584278 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b911b49-7b0f-48ef-9626-cd43d308d596" containerName="oc" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.584456 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc410de-5b42-44d1-8b29-37161475730e" containerName="neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.584476 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b911b49-7b0f-48ef-9626-cd43d308d596" containerName="oc" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.585066 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.587903 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.588054 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.588121 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.589161 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.591065 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.624228 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8"] Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.763365 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.763427 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.763700 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.763769 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42btt\" (UniqueName: \"kubernetes.io/projected/8713b951-b516-42bd-9286-4343e5bcc955-kube-api-access-42btt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.763803 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.868379 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.868450 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.868531 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.868555 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42btt\" (UniqueName: \"kubernetes.io/projected/8713b951-b516-42bd-9286-4343e5bcc955-kube-api-access-42btt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.868578 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.873451 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.873901 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.877032 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.882977 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.893893 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42btt\" (UniqueName: \"kubernetes.io/projected/8713b951-b516-42bd-9286-4343e5bcc955-kube-api-access-42btt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:53 crc kubenswrapper[4778]: I0312 13:50:53.907570 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:50:54 crc kubenswrapper[4778]: I0312 13:50:54.450962 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8"] Mar 12 13:50:55 crc kubenswrapper[4778]: I0312 13:50:55.420237 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" event={"ID":"8713b951-b516-42bd-9286-4343e5bcc955","Type":"ContainerStarted","Data":"a4247422875ccc1a942b0fe9bbe7105ae4cc94d3cc420523932f239263b637f2"} Mar 12 13:50:55 crc kubenswrapper[4778]: I0312 13:50:55.420824 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" event={"ID":"8713b951-b516-42bd-9286-4343e5bcc955","Type":"ContainerStarted","Data":"cceb4d2cb4de9b629a148def19221cfb5724a46254ab30fcc858bc4f7a667a5f"} Mar 12 13:50:55 crc kubenswrapper[4778]: I0312 13:50:55.451022 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" podStartSLOduration=1.968408094 podStartE2EDuration="2.450997943s" podCreationTimestamp="2026-03-12 13:50:53 +0000 UTC" firstStartedPulling="2026-03-12 13:50:54.452425869 +0000 UTC m=+2472.901121265" lastFinishedPulling="2026-03-12 13:50:54.935015718 +0000 UTC m=+2473.383711114" observedRunningTime="2026-03-12 13:50:55.440501765 +0000 UTC m=+2473.889197171" watchObservedRunningTime="2026-03-12 13:50:55.450997943 +0000 UTC m=+2473.899693339" Mar 12 13:50:58 crc kubenswrapper[4778]: I0312 13:50:58.254522 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:50:58 crc kubenswrapper[4778]: E0312 13:50:58.255053 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:51:09 crc kubenswrapper[4778]: I0312 13:51:09.254114 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:51:09 crc kubenswrapper[4778]: E0312 13:51:09.256764 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:51:23 crc kubenswrapper[4778]: I0312 13:51:23.253920 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:51:23 crc kubenswrapper[4778]: E0312 13:51:23.254706 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:51:38 crc kubenswrapper[4778]: I0312 13:51:38.253783 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:51:38 crc kubenswrapper[4778]: E0312 13:51:38.255564 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:51:50 crc kubenswrapper[4778]: I0312 13:51:50.274388 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:51:50 crc kubenswrapper[4778]: E0312 13:51:50.275296 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.176490 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555392-wg78w"] Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.179136 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.181618 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.182497 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.183945 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.197161 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjqkg\" (UniqueName: \"kubernetes.io/projected/0a0b5070-03d8-45fe-8148-c39a9b560fbb-kube-api-access-hjqkg\") pod \"auto-csr-approver-29555392-wg78w\" (UID: \"0a0b5070-03d8-45fe-8148-c39a9b560fbb\") " pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.200394 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555392-wg78w"] Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.298974 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjqkg\" (UniqueName: \"kubernetes.io/projected/0a0b5070-03d8-45fe-8148-c39a9b560fbb-kube-api-access-hjqkg\") pod \"auto-csr-approver-29555392-wg78w\" (UID: \"0a0b5070-03d8-45fe-8148-c39a9b560fbb\") " pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.323907 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjqkg\" (UniqueName: \"kubernetes.io/projected/0a0b5070-03d8-45fe-8148-c39a9b560fbb-kube-api-access-hjqkg\") pod \"auto-csr-approver-29555392-wg78w\" (UID: \"0a0b5070-03d8-45fe-8148-c39a9b560fbb\") " pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.503749 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:00 crc kubenswrapper[4778]: I0312 13:52:00.952958 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555392-wg78w"] Mar 12 13:52:01 crc kubenswrapper[4778]: I0312 13:52:01.120319 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555392-wg78w" event={"ID":"0a0b5070-03d8-45fe-8148-c39a9b560fbb","Type":"ContainerStarted","Data":"754978825e789f510db4cf26367293fc9868be6072bc39dc3e4bce6fab2beb06"} Mar 12 13:52:04 crc kubenswrapper[4778]: I0312 13:52:04.254271 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:52:04 crc kubenswrapper[4778]: E0312 13:52:04.255057 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:52:05 crc kubenswrapper[4778]: I0312 13:52:05.184531 4778 generic.go:334] "Generic (PLEG): container finished" podID="0a0b5070-03d8-45fe-8148-c39a9b560fbb" containerID="1f9b06fe647c9c9d52674fc3e58e1c9d5c930036da2b4f235a350fc83217496f" exitCode=0 Mar 12 13:52:05 crc kubenswrapper[4778]: I0312 13:52:05.184818 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555392-wg78w" event={"ID":"0a0b5070-03d8-45fe-8148-c39a9b560fbb","Type":"ContainerDied","Data":"1f9b06fe647c9c9d52674fc3e58e1c9d5c930036da2b4f235a350fc83217496f"} Mar 12 13:52:06 crc kubenswrapper[4778]: I0312 13:52:06.505048 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:06 crc kubenswrapper[4778]: I0312 13:52:06.648350 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjqkg\" (UniqueName: \"kubernetes.io/projected/0a0b5070-03d8-45fe-8148-c39a9b560fbb-kube-api-access-hjqkg\") pod \"0a0b5070-03d8-45fe-8148-c39a9b560fbb\" (UID: \"0a0b5070-03d8-45fe-8148-c39a9b560fbb\") " Mar 12 13:52:06 crc kubenswrapper[4778]: I0312 13:52:06.657216 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a0b5070-03d8-45fe-8148-c39a9b560fbb-kube-api-access-hjqkg" (OuterVolumeSpecName: "kube-api-access-hjqkg") pod "0a0b5070-03d8-45fe-8148-c39a9b560fbb" (UID: "0a0b5070-03d8-45fe-8148-c39a9b560fbb"). InnerVolumeSpecName "kube-api-access-hjqkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:52:06 crc kubenswrapper[4778]: I0312 13:52:06.750843 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjqkg\" (UniqueName: \"kubernetes.io/projected/0a0b5070-03d8-45fe-8148-c39a9b560fbb-kube-api-access-hjqkg\") on node \"crc\" DevicePath \"\"" Mar 12 13:52:07 crc kubenswrapper[4778]: I0312 13:52:07.202568 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555392-wg78w" event={"ID":"0a0b5070-03d8-45fe-8148-c39a9b560fbb","Type":"ContainerDied","Data":"754978825e789f510db4cf26367293fc9868be6072bc39dc3e4bce6fab2beb06"} Mar 12 13:52:07 crc kubenswrapper[4778]: I0312 13:52:07.202632 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="754978825e789f510db4cf26367293fc9868be6072bc39dc3e4bce6fab2beb06" Mar 12 13:52:07 crc kubenswrapper[4778]: I0312 13:52:07.202642 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555392-wg78w" Mar 12 13:52:07 crc kubenswrapper[4778]: I0312 13:52:07.580305 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555386-vjswk"] Mar 12 13:52:07 crc kubenswrapper[4778]: I0312 13:52:07.588654 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555386-vjswk"] Mar 12 13:52:08 crc kubenswrapper[4778]: I0312 13:52:08.267583 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6" path="/var/lib/kubelet/pods/f55c85e9-4cb7-4ac4-bc3d-c37217b4abf6/volumes" Mar 12 13:52:15 crc kubenswrapper[4778]: I0312 13:52:15.253902 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:52:15 crc kubenswrapper[4778]: E0312 13:52:15.254769 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:52:19 crc kubenswrapper[4778]: I0312 13:52:19.713797 4778 scope.go:117] "RemoveContainer" containerID="2e424e585231dad361491fa12a9a787f83d6973879b6b45159764198bbcf5877" Mar 12 13:52:26 crc kubenswrapper[4778]: I0312 13:52:26.253808 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:52:26 crc kubenswrapper[4778]: E0312 13:52:26.254534 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:52:40 crc kubenswrapper[4778]: I0312 13:52:40.254327 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:52:40 crc kubenswrapper[4778]: E0312 13:52:40.255473 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:52:54 crc kubenswrapper[4778]: I0312 13:52:54.253724 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:52:54 crc kubenswrapper[4778]: E0312 13:52:54.254434 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:53:09 crc kubenswrapper[4778]: I0312 13:53:09.253986 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:53:09 crc kubenswrapper[4778]: I0312 13:53:09.738404 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"e20e6fa2d381e3ff917a0f6074e27521c909a7932045eacfc15c005ed843cb93"} Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.150542 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555394-7f7nf"] Mar 12 13:54:00 crc kubenswrapper[4778]: E0312 13:54:00.151578 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a0b5070-03d8-45fe-8148-c39a9b560fbb" containerName="oc" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.151594 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a0b5070-03d8-45fe-8148-c39a9b560fbb" containerName="oc" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.151837 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a0b5070-03d8-45fe-8148-c39a9b560fbb" containerName="oc" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.152634 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.154694 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.154753 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.155420 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.162292 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555394-7f7nf"] Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.220290 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww9dv\" (UniqueName: \"kubernetes.io/projected/65fbb68a-57a2-40bf-9149-6cfe13fe147c-kube-api-access-ww9dv\") pod \"auto-csr-approver-29555394-7f7nf\" (UID: \"65fbb68a-57a2-40bf-9149-6cfe13fe147c\") " pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.322799 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww9dv\" (UniqueName: \"kubernetes.io/projected/65fbb68a-57a2-40bf-9149-6cfe13fe147c-kube-api-access-ww9dv\") pod \"auto-csr-approver-29555394-7f7nf\" (UID: \"65fbb68a-57a2-40bf-9149-6cfe13fe147c\") " pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.349125 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww9dv\" (UniqueName: \"kubernetes.io/projected/65fbb68a-57a2-40bf-9149-6cfe13fe147c-kube-api-access-ww9dv\") pod \"auto-csr-approver-29555394-7f7nf\" (UID: \"65fbb68a-57a2-40bf-9149-6cfe13fe147c\") " pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.479548 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:00 crc kubenswrapper[4778]: I0312 13:54:00.932120 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555394-7f7nf"] Mar 12 13:54:00 crc kubenswrapper[4778]: W0312 13:54:00.938485 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65fbb68a_57a2_40bf_9149_6cfe13fe147c.slice/crio-0cb9aa3c58c77f6af33863cdd5012dbf15ab7d672bd63a4b37baf0edea4c0df3 WatchSource:0}: Error finding container 0cb9aa3c58c77f6af33863cdd5012dbf15ab7d672bd63a4b37baf0edea4c0df3: Status 404 returned error can't find the container with id 0cb9aa3c58c77f6af33863cdd5012dbf15ab7d672bd63a4b37baf0edea4c0df3 Mar 12 13:54:01 crc kubenswrapper[4778]: I0312 13:54:01.172423 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" event={"ID":"65fbb68a-57a2-40bf-9149-6cfe13fe147c","Type":"ContainerStarted","Data":"0cb9aa3c58c77f6af33863cdd5012dbf15ab7d672bd63a4b37baf0edea4c0df3"} Mar 12 13:54:03 crc kubenswrapper[4778]: I0312 13:54:03.192373 4778 generic.go:334] "Generic (PLEG): container finished" podID="65fbb68a-57a2-40bf-9149-6cfe13fe147c" containerID="b93a8a130b5f9b7d0852157c6942677a4b8f445ae1cc7062b429977ab9491779" exitCode=0 Mar 12 13:54:03 crc kubenswrapper[4778]: I0312 13:54:03.192449 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" event={"ID":"65fbb68a-57a2-40bf-9149-6cfe13fe147c","Type":"ContainerDied","Data":"b93a8a130b5f9b7d0852157c6942677a4b8f445ae1cc7062b429977ab9491779"} Mar 12 13:54:04 crc kubenswrapper[4778]: I0312 13:54:04.560559 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:04 crc kubenswrapper[4778]: I0312 13:54:04.706091 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww9dv\" (UniqueName: \"kubernetes.io/projected/65fbb68a-57a2-40bf-9149-6cfe13fe147c-kube-api-access-ww9dv\") pod \"65fbb68a-57a2-40bf-9149-6cfe13fe147c\" (UID: \"65fbb68a-57a2-40bf-9149-6cfe13fe147c\") " Mar 12 13:54:04 crc kubenswrapper[4778]: I0312 13:54:04.711896 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65fbb68a-57a2-40bf-9149-6cfe13fe147c-kube-api-access-ww9dv" (OuterVolumeSpecName: "kube-api-access-ww9dv") pod "65fbb68a-57a2-40bf-9149-6cfe13fe147c" (UID: "65fbb68a-57a2-40bf-9149-6cfe13fe147c"). InnerVolumeSpecName "kube-api-access-ww9dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:54:04 crc kubenswrapper[4778]: I0312 13:54:04.808306 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww9dv\" (UniqueName: \"kubernetes.io/projected/65fbb68a-57a2-40bf-9149-6cfe13fe147c-kube-api-access-ww9dv\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:05 crc kubenswrapper[4778]: I0312 13:54:05.214033 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" event={"ID":"65fbb68a-57a2-40bf-9149-6cfe13fe147c","Type":"ContainerDied","Data":"0cb9aa3c58c77f6af33863cdd5012dbf15ab7d672bd63a4b37baf0edea4c0df3"} Mar 12 13:54:05 crc kubenswrapper[4778]: I0312 13:54:05.214581 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cb9aa3c58c77f6af33863cdd5012dbf15ab7d672bd63a4b37baf0edea4c0df3" Mar 12 13:54:05 crc kubenswrapper[4778]: I0312 13:54:05.214094 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555394-7f7nf" Mar 12 13:54:05 crc kubenswrapper[4778]: I0312 13:54:05.630875 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555388-5mnjh"] Mar 12 13:54:05 crc kubenswrapper[4778]: I0312 13:54:05.638892 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555388-5mnjh"] Mar 12 13:54:06 crc kubenswrapper[4778]: I0312 13:54:06.265742 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="110071e6-5231-434c-af16-87b68a3d0c8f" path="/var/lib/kubelet/pods/110071e6-5231-434c-af16-87b68a3d0c8f/volumes" Mar 12 13:54:19 crc kubenswrapper[4778]: I0312 13:54:19.846047 4778 scope.go:117] "RemoveContainer" containerID="2b6df075041c6c1583e329716aacbd0c53d1a64cadc9905cc6ddb1e1bd9b676d" Mar 12 13:54:40 crc kubenswrapper[4778]: I0312 13:54:40.972078 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dr7tv"] Mar 12 13:54:40 crc kubenswrapper[4778]: E0312 13:54:40.973327 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fbb68a-57a2-40bf-9149-6cfe13fe147c" containerName="oc" Mar 12 13:54:40 crc kubenswrapper[4778]: I0312 13:54:40.973355 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fbb68a-57a2-40bf-9149-6cfe13fe147c" containerName="oc" Mar 12 13:54:40 crc kubenswrapper[4778]: I0312 13:54:40.973612 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="65fbb68a-57a2-40bf-9149-6cfe13fe147c" containerName="oc" Mar 12 13:54:40 crc kubenswrapper[4778]: I0312 13:54:40.975556 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:40 crc kubenswrapper[4778]: I0312 13:54:40.985126 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dr7tv"] Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.151666 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdhb\" (UniqueName: \"kubernetes.io/projected/2175642e-200d-49e4-b07c-e594a50dec28-kube-api-access-zxdhb\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.151883 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-utilities\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.151923 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-catalog-content\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.254069 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdhb\" (UniqueName: \"kubernetes.io/projected/2175642e-200d-49e4-b07c-e594a50dec28-kube-api-access-zxdhb\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.254279 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-utilities\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.254314 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-catalog-content\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.254848 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-catalog-content\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.254859 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-utilities\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.283337 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdhb\" (UniqueName: \"kubernetes.io/projected/2175642e-200d-49e4-b07c-e594a50dec28-kube-api-access-zxdhb\") pod \"certified-operators-dr7tv\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.298996 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:41 crc kubenswrapper[4778]: I0312 13:54:41.972821 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dr7tv"] Mar 12 13:54:42 crc kubenswrapper[4778]: I0312 13:54:42.527683 4778 generic.go:334] "Generic (PLEG): container finished" podID="2175642e-200d-49e4-b07c-e594a50dec28" containerID="78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3" exitCode=0 Mar 12 13:54:42 crc kubenswrapper[4778]: I0312 13:54:42.527746 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerDied","Data":"78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3"} Mar 12 13:54:42 crc kubenswrapper[4778]: I0312 13:54:42.527932 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerStarted","Data":"4b66228125468deb0c32f8bf58c76651899d1e61cf9f3d4b448151fde53f1bc3"} Mar 12 13:54:43 crc kubenswrapper[4778]: I0312 13:54:43.546209 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerStarted","Data":"a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46"} Mar 12 13:54:44 crc kubenswrapper[4778]: I0312 13:54:44.563529 4778 generic.go:334] "Generic (PLEG): container finished" podID="2175642e-200d-49e4-b07c-e594a50dec28" containerID="a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46" exitCode=0 Mar 12 13:54:44 crc kubenswrapper[4778]: I0312 13:54:44.563581 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerDied","Data":"a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46"} Mar 12 13:54:45 crc kubenswrapper[4778]: I0312 13:54:45.573826 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerStarted","Data":"0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70"} Mar 12 13:54:45 crc kubenswrapper[4778]: I0312 13:54:45.603635 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dr7tv" podStartSLOduration=3.132173988 podStartE2EDuration="5.603611248s" podCreationTimestamp="2026-03-12 13:54:40 +0000 UTC" firstStartedPulling="2026-03-12 13:54:42.529545341 +0000 UTC m=+2700.978240737" lastFinishedPulling="2026-03-12 13:54:45.000982581 +0000 UTC m=+2703.449677997" observedRunningTime="2026-03-12 13:54:45.597212816 +0000 UTC m=+2704.045908232" watchObservedRunningTime="2026-03-12 13:54:45.603611248 +0000 UTC m=+2704.052306644" Mar 12 13:54:46 crc kubenswrapper[4778]: I0312 13:54:46.952011 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gq2rs"] Mar 12 13:54:46 crc kubenswrapper[4778]: I0312 13:54:46.954558 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:46 crc kubenswrapper[4778]: I0312 13:54:46.970357 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gq2rs"] Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.060930 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-catalog-content\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.061053 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-utilities\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.061082 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt2rb\" (UniqueName: \"kubernetes.io/projected/1408ca57-ea2c-414f-93af-8c6b930c1fe4-kube-api-access-xt2rb\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.162944 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-catalog-content\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.163067 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-utilities\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.163093 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt2rb\" (UniqueName: \"kubernetes.io/projected/1408ca57-ea2c-414f-93af-8c6b930c1fe4-kube-api-access-xt2rb\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.163587 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-catalog-content\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.163623 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-utilities\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.186227 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt2rb\" (UniqueName: \"kubernetes.io/projected/1408ca57-ea2c-414f-93af-8c6b930c1fe4-kube-api-access-xt2rb\") pod \"community-operators-gq2rs\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.302032 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:47 crc kubenswrapper[4778]: I0312 13:54:47.874743 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gq2rs"] Mar 12 13:54:47 crc kubenswrapper[4778]: W0312 13:54:47.875724 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1408ca57_ea2c_414f_93af_8c6b930c1fe4.slice/crio-e3dfe06f339548a2b7f51a541535fbe107a48486dc81c829d0a16d0d9b1044de WatchSource:0}: Error finding container e3dfe06f339548a2b7f51a541535fbe107a48486dc81c829d0a16d0d9b1044de: Status 404 returned error can't find the container with id e3dfe06f339548a2b7f51a541535fbe107a48486dc81c829d0a16d0d9b1044de Mar 12 13:54:48 crc kubenswrapper[4778]: I0312 13:54:48.600650 4778 generic.go:334] "Generic (PLEG): container finished" podID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerID="bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a" exitCode=0 Mar 12 13:54:48 crc kubenswrapper[4778]: I0312 13:54:48.601070 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerDied","Data":"bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a"} Mar 12 13:54:48 crc kubenswrapper[4778]: I0312 13:54:48.601100 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerStarted","Data":"e3dfe06f339548a2b7f51a541535fbe107a48486dc81c829d0a16d0d9b1044de"} Mar 12 13:54:48 crc kubenswrapper[4778]: I0312 13:54:48.617162 4778 generic.go:334] "Generic (PLEG): container finished" podID="8713b951-b516-42bd-9286-4343e5bcc955" containerID="a4247422875ccc1a942b0fe9bbe7105ae4cc94d3cc420523932f239263b637f2" exitCode=0 Mar 12 13:54:48 crc kubenswrapper[4778]: I0312 13:54:48.617287 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" event={"ID":"8713b951-b516-42bd-9286-4343e5bcc955","Type":"ContainerDied","Data":"a4247422875ccc1a942b0fe9bbe7105ae4cc94d3cc420523932f239263b637f2"} Mar 12 13:54:49 crc kubenswrapper[4778]: I0312 13:54:49.627773 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerStarted","Data":"da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c"} Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.073283 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.223653 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42btt\" (UniqueName: \"kubernetes.io/projected/8713b951-b516-42bd-9286-4343e5bcc955-kube-api-access-42btt\") pod \"8713b951-b516-42bd-9286-4343e5bcc955\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.223969 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-inventory\") pod \"8713b951-b516-42bd-9286-4343e5bcc955\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.224031 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-ssh-key-openstack-edpm-ipam\") pod \"8713b951-b516-42bd-9286-4343e5bcc955\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.224065 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-combined-ca-bundle\") pod \"8713b951-b516-42bd-9286-4343e5bcc955\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.224146 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-secret-0\") pod \"8713b951-b516-42bd-9286-4343e5bcc955\" (UID: \"8713b951-b516-42bd-9286-4343e5bcc955\") " Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.228973 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8713b951-b516-42bd-9286-4343e5bcc955" (UID: "8713b951-b516-42bd-9286-4343e5bcc955"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.237683 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8713b951-b516-42bd-9286-4343e5bcc955-kube-api-access-42btt" (OuterVolumeSpecName: "kube-api-access-42btt") pod "8713b951-b516-42bd-9286-4343e5bcc955" (UID: "8713b951-b516-42bd-9286-4343e5bcc955"). InnerVolumeSpecName "kube-api-access-42btt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.252379 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8713b951-b516-42bd-9286-4343e5bcc955" (UID: "8713b951-b516-42bd-9286-4343e5bcc955"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.254366 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-inventory" (OuterVolumeSpecName: "inventory") pod "8713b951-b516-42bd-9286-4343e5bcc955" (UID: "8713b951-b516-42bd-9286-4343e5bcc955"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.268377 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "8713b951-b516-42bd-9286-4343e5bcc955" (UID: "8713b951-b516-42bd-9286-4343e5bcc955"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.326502 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.326542 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.326555 4778 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.326565 4778 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8713b951-b516-42bd-9286-4343e5bcc955-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.326573 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42btt\" (UniqueName: \"kubernetes.io/projected/8713b951-b516-42bd-9286-4343e5bcc955-kube-api-access-42btt\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.636259 4778 generic.go:334] "Generic (PLEG): container finished" podID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerID="da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c" exitCode=0 Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.636301 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerDied","Data":"da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c"} Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.637503 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" event={"ID":"8713b951-b516-42bd-9286-4343e5bcc955","Type":"ContainerDied","Data":"cceb4d2cb4de9b629a148def19221cfb5724a46254ab30fcc858bc4f7a667a5f"} Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.637536 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cceb4d2cb4de9b629a148def19221cfb5724a46254ab30fcc858bc4f7a667a5f" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.637553 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.739614 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s"] Mar 12 13:54:50 crc kubenswrapper[4778]: E0312 13:54:50.740050 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8713b951-b516-42bd-9286-4343e5bcc955" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.740066 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8713b951-b516-42bd-9286-4343e5bcc955" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.740248 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="8713b951-b516-42bd-9286-4343e5bcc955" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.740881 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.744654 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.744685 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.744781 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.745276 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.745320 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.746808 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.749876 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.757593 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s"] Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.835238 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.835552 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.835660 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.835854 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.835969 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.836070 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.836156 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.836397 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.836496 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.836533 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp6ls\" (UniqueName: \"kubernetes.io/projected/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-kube-api-access-fp6ls\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.836593 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.938722 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939423 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939457 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939473 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939500 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939526 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939546 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp6ls\" (UniqueName: \"kubernetes.io/projected/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-kube-api-access-fp6ls\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939579 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939611 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939668 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.939689 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.940733 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.943728 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.944252 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.945742 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.946935 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.947243 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.949641 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.950833 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.952505 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.961601 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:50 crc kubenswrapper[4778]: I0312 13:54:50.963334 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp6ls\" (UniqueName: \"kubernetes.io/projected/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-kube-api-access-fp6ls\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5tw6s\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.062245 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.299660 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.299704 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.366475 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.582668 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s"] Mar 12 13:54:51 crc kubenswrapper[4778]: W0312 13:54:51.587638 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ed77f87_e6b2_4c7a_8b0e_003106200dc8.slice/crio-1617b96c98df28869b5a069f5b74bb8126ce4a98898565a1251dadc01020d162 WatchSource:0}: Error finding container 1617b96c98df28869b5a069f5b74bb8126ce4a98898565a1251dadc01020d162: Status 404 returned error can't find the container with id 1617b96c98df28869b5a069f5b74bb8126ce4a98898565a1251dadc01020d162 Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.648240 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerStarted","Data":"11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00"} Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.649949 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" event={"ID":"6ed77f87-e6b2-4c7a-8b0e-003106200dc8","Type":"ContainerStarted","Data":"1617b96c98df28869b5a069f5b74bb8126ce4a98898565a1251dadc01020d162"} Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.671046 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gq2rs" podStartSLOduration=2.916142089 podStartE2EDuration="5.67102827s" podCreationTimestamp="2026-03-12 13:54:46 +0000 UTC" firstStartedPulling="2026-03-12 13:54:48.609215401 +0000 UTC m=+2707.057910797" lastFinishedPulling="2026-03-12 13:54:51.364101582 +0000 UTC m=+2709.812796978" observedRunningTime="2026-03-12 13:54:51.666291926 +0000 UTC m=+2710.114987322" watchObservedRunningTime="2026-03-12 13:54:51.67102827 +0000 UTC m=+2710.119723666" Mar 12 13:54:51 crc kubenswrapper[4778]: I0312 13:54:51.699520 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:52 crc kubenswrapper[4778]: I0312 13:54:52.674312 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" event={"ID":"6ed77f87-e6b2-4c7a-8b0e-003106200dc8","Type":"ContainerStarted","Data":"e2a35e751ce79cb5226fd46ca73472f5cd7c47201c7bf749d5ffd3dae25fcc72"} Mar 12 13:54:52 crc kubenswrapper[4778]: I0312 13:54:52.696858 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" podStartSLOduration=1.943705015 podStartE2EDuration="2.696837788s" podCreationTimestamp="2026-03-12 13:54:50 +0000 UTC" firstStartedPulling="2026-03-12 13:54:51.592638143 +0000 UTC m=+2710.041333539" lastFinishedPulling="2026-03-12 13:54:52.345770916 +0000 UTC m=+2710.794466312" observedRunningTime="2026-03-12 13:54:52.695590843 +0000 UTC m=+2711.144286259" watchObservedRunningTime="2026-03-12 13:54:52.696837788 +0000 UTC m=+2711.145533194" Mar 12 13:54:53 crc kubenswrapper[4778]: I0312 13:54:53.742036 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dr7tv"] Mar 12 13:54:53 crc kubenswrapper[4778]: I0312 13:54:53.742302 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dr7tv" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="registry-server" containerID="cri-o://0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70" gracePeriod=2 Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.680395 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.712537 4778 generic.go:334] "Generic (PLEG): container finished" podID="2175642e-200d-49e4-b07c-e594a50dec28" containerID="0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70" exitCode=0 Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.712931 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerDied","Data":"0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70"} Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.712981 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dr7tv" event={"ID":"2175642e-200d-49e4-b07c-e594a50dec28","Type":"ContainerDied","Data":"4b66228125468deb0c32f8bf58c76651899d1e61cf9f3d4b448151fde53f1bc3"} Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.713035 4778 scope.go:117] "RemoveContainer" containerID="0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.713414 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dr7tv" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.755147 4778 scope.go:117] "RemoveContainer" containerID="a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.774590 4778 scope.go:117] "RemoveContainer" containerID="78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.819948 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxdhb\" (UniqueName: \"kubernetes.io/projected/2175642e-200d-49e4-b07c-e594a50dec28-kube-api-access-zxdhb\") pod \"2175642e-200d-49e4-b07c-e594a50dec28\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.820145 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-utilities\") pod \"2175642e-200d-49e4-b07c-e594a50dec28\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.820257 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-catalog-content\") pod \"2175642e-200d-49e4-b07c-e594a50dec28\" (UID: \"2175642e-200d-49e4-b07c-e594a50dec28\") " Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.823158 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-utilities" (OuterVolumeSpecName: "utilities") pod "2175642e-200d-49e4-b07c-e594a50dec28" (UID: "2175642e-200d-49e4-b07c-e594a50dec28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.829734 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2175642e-200d-49e4-b07c-e594a50dec28-kube-api-access-zxdhb" (OuterVolumeSpecName: "kube-api-access-zxdhb") pod "2175642e-200d-49e4-b07c-e594a50dec28" (UID: "2175642e-200d-49e4-b07c-e594a50dec28"). InnerVolumeSpecName "kube-api-access-zxdhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.833625 4778 scope.go:117] "RemoveContainer" containerID="0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70" Mar 12 13:54:54 crc kubenswrapper[4778]: E0312 13:54:54.834287 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70\": container with ID starting with 0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70 not found: ID does not exist" containerID="0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.834339 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70"} err="failed to get container status \"0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70\": rpc error: code = NotFound desc = could not find container \"0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70\": container with ID starting with 0205168eb943737b001c1df62f66ef855d9c4f763c3adaea0c470e9c1ddc2f70 not found: ID does not exist" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.834369 4778 scope.go:117] "RemoveContainer" containerID="a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46" Mar 12 13:54:54 crc kubenswrapper[4778]: E0312 13:54:54.834863 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46\": container with ID starting with a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46 not found: ID does not exist" containerID="a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.834916 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46"} err="failed to get container status \"a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46\": rpc error: code = NotFound desc = could not find container \"a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46\": container with ID starting with a9205572a7471ec7794f4549cb1ca055cf820789a0c4982c3e93ca87112c9e46 not found: ID does not exist" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.834949 4778 scope.go:117] "RemoveContainer" containerID="78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3" Mar 12 13:54:54 crc kubenswrapper[4778]: E0312 13:54:54.835466 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3\": container with ID starting with 78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3 not found: ID does not exist" containerID="78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.835505 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3"} err="failed to get container status \"78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3\": rpc error: code = NotFound desc = could not find container \"78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3\": container with ID starting with 78f31326cd9cdc9aaf40af7d920b0b345a573b27f1b4f379a64419fc772025c3 not found: ID does not exist" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.894592 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2175642e-200d-49e4-b07c-e594a50dec28" (UID: "2175642e-200d-49e4-b07c-e594a50dec28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.923819 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.923858 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2175642e-200d-49e4-b07c-e594a50dec28-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:54 crc kubenswrapper[4778]: I0312 13:54:54.923871 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxdhb\" (UniqueName: \"kubernetes.io/projected/2175642e-200d-49e4-b07c-e594a50dec28-kube-api-access-zxdhb\") on node \"crc\" DevicePath \"\"" Mar 12 13:54:55 crc kubenswrapper[4778]: I0312 13:54:55.052976 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dr7tv"] Mar 12 13:54:55 crc kubenswrapper[4778]: I0312 13:54:55.061288 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dr7tv"] Mar 12 13:54:56 crc kubenswrapper[4778]: I0312 13:54:56.271699 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2175642e-200d-49e4-b07c-e594a50dec28" path="/var/lib/kubelet/pods/2175642e-200d-49e4-b07c-e594a50dec28/volumes" Mar 12 13:54:57 crc kubenswrapper[4778]: I0312 13:54:57.302801 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:57 crc kubenswrapper[4778]: I0312 13:54:57.303138 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:57 crc kubenswrapper[4778]: I0312 13:54:57.357548 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:57 crc kubenswrapper[4778]: I0312 13:54:57.815452 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:54:58 crc kubenswrapper[4778]: I0312 13:54:58.751490 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gq2rs"] Mar 12 13:54:59 crc kubenswrapper[4778]: I0312 13:54:59.767562 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gq2rs" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="registry-server" containerID="cri-o://11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00" gracePeriod=2 Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.244945 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.338918 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xt2rb\" (UniqueName: \"kubernetes.io/projected/1408ca57-ea2c-414f-93af-8c6b930c1fe4-kube-api-access-xt2rb\") pod \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.339242 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-utilities\") pod \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.339311 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-catalog-content\") pod \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\" (UID: \"1408ca57-ea2c-414f-93af-8c6b930c1fe4\") " Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.340308 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-utilities" (OuterVolumeSpecName: "utilities") pod "1408ca57-ea2c-414f-93af-8c6b930c1fe4" (UID: "1408ca57-ea2c-414f-93af-8c6b930c1fe4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.344958 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1408ca57-ea2c-414f-93af-8c6b930c1fe4-kube-api-access-xt2rb" (OuterVolumeSpecName: "kube-api-access-xt2rb") pod "1408ca57-ea2c-414f-93af-8c6b930c1fe4" (UID: "1408ca57-ea2c-414f-93af-8c6b930c1fe4"). InnerVolumeSpecName "kube-api-access-xt2rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.442008 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.442051 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xt2rb\" (UniqueName: \"kubernetes.io/projected/1408ca57-ea2c-414f-93af-8c6b930c1fe4-kube-api-access-xt2rb\") on node \"crc\" DevicePath \"\"" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.466279 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1408ca57-ea2c-414f-93af-8c6b930c1fe4" (UID: "1408ca57-ea2c-414f-93af-8c6b930c1fe4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.545259 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1408ca57-ea2c-414f-93af-8c6b930c1fe4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.777515 4778 generic.go:334] "Generic (PLEG): container finished" podID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerID="11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00" exitCode=0 Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.777557 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerDied","Data":"11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00"} Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.777582 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gq2rs" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.777607 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gq2rs" event={"ID":"1408ca57-ea2c-414f-93af-8c6b930c1fe4","Type":"ContainerDied","Data":"e3dfe06f339548a2b7f51a541535fbe107a48486dc81c829d0a16d0d9b1044de"} Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.777630 4778 scope.go:117] "RemoveContainer" containerID="11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.821144 4778 scope.go:117] "RemoveContainer" containerID="da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.823749 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gq2rs"] Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.832047 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gq2rs"] Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.845906 4778 scope.go:117] "RemoveContainer" containerID="bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.885714 4778 scope.go:117] "RemoveContainer" containerID="11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00" Mar 12 13:55:00 crc kubenswrapper[4778]: E0312 13:55:00.886159 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00\": container with ID starting with 11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00 not found: ID does not exist" containerID="11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.886313 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00"} err="failed to get container status \"11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00\": rpc error: code = NotFound desc = could not find container \"11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00\": container with ID starting with 11a3098d454804d45f104a9f35c5a7438c4227caaaef5bf8fde9f94039e4ae00 not found: ID does not exist" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.886340 4778 scope.go:117] "RemoveContainer" containerID="da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c" Mar 12 13:55:00 crc kubenswrapper[4778]: E0312 13:55:00.886660 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c\": container with ID starting with da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c not found: ID does not exist" containerID="da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.886711 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c"} err="failed to get container status \"da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c\": rpc error: code = NotFound desc = could not find container \"da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c\": container with ID starting with da490793edbc377cb06911d757533645c04fa6ea49dc72e732862ebbe646049c not found: ID does not exist" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.886737 4778 scope.go:117] "RemoveContainer" containerID="bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a" Mar 12 13:55:00 crc kubenswrapper[4778]: E0312 13:55:00.887006 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a\": container with ID starting with bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a not found: ID does not exist" containerID="bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a" Mar 12 13:55:00 crc kubenswrapper[4778]: I0312 13:55:00.887032 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a"} err="failed to get container status \"bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a\": rpc error: code = NotFound desc = could not find container \"bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a\": container with ID starting with bebe7cef5bff75eca18b4a61abf22de1f6ce00043a7457a97f1dd61e0e7a224a not found: ID does not exist" Mar 12 13:55:02 crc kubenswrapper[4778]: I0312 13:55:02.265232 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" path="/var/lib/kubelet/pods/1408ca57-ea2c-414f-93af-8c6b930c1fe4/volumes" Mar 12 13:55:28 crc kubenswrapper[4778]: I0312 13:55:28.557946 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:55:28 crc kubenswrapper[4778]: I0312 13:55:28.558469 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:55:58 crc kubenswrapper[4778]: I0312 13:55:58.557727 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:55:58 crc kubenswrapper[4778]: I0312 13:55:58.558373 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.147858 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555396-lhqkd"] Mar 12 13:56:00 crc kubenswrapper[4778]: E0312 13:56:00.148563 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="extract-utilities" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148575 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="extract-utilities" Mar 12 13:56:00 crc kubenswrapper[4778]: E0312 13:56:00.148587 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="extract-content" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148592 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="extract-content" Mar 12 13:56:00 crc kubenswrapper[4778]: E0312 13:56:00.148602 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="registry-server" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148609 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="registry-server" Mar 12 13:56:00 crc kubenswrapper[4778]: E0312 13:56:00.148619 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="extract-utilities" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148625 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="extract-utilities" Mar 12 13:56:00 crc kubenswrapper[4778]: E0312 13:56:00.148633 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="registry-server" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148638 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="registry-server" Mar 12 13:56:00 crc kubenswrapper[4778]: E0312 13:56:00.148646 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="extract-content" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148652 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="extract-content" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148826 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1408ca57-ea2c-414f-93af-8c6b930c1fe4" containerName="registry-server" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.148837 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2175642e-200d-49e4-b07c-e594a50dec28" containerName="registry-server" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.149485 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.151732 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.151869 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.152087 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.166216 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555396-lhqkd"] Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.228356 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59ctl\" (UniqueName: \"kubernetes.io/projected/90b32527-d7b2-4938-a8c2-882067947e78-kube-api-access-59ctl\") pod \"auto-csr-approver-29555396-lhqkd\" (UID: \"90b32527-d7b2-4938-a8c2-882067947e78\") " pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.329875 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59ctl\" (UniqueName: \"kubernetes.io/projected/90b32527-d7b2-4938-a8c2-882067947e78-kube-api-access-59ctl\") pod \"auto-csr-approver-29555396-lhqkd\" (UID: \"90b32527-d7b2-4938-a8c2-882067947e78\") " pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.349504 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59ctl\" (UniqueName: \"kubernetes.io/projected/90b32527-d7b2-4938-a8c2-882067947e78-kube-api-access-59ctl\") pod \"auto-csr-approver-29555396-lhqkd\" (UID: \"90b32527-d7b2-4938-a8c2-882067947e78\") " pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.472171 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.927452 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 13:56:00 crc kubenswrapper[4778]: I0312 13:56:00.927484 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555396-lhqkd"] Mar 12 13:56:01 crc kubenswrapper[4778]: I0312 13:56:01.345819 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" event={"ID":"90b32527-d7b2-4938-a8c2-882067947e78","Type":"ContainerStarted","Data":"4fba4a78ca2fd2f35531ebdaf442cdd6b4b4c347b21e38401f16971eae6edcc0"} Mar 12 13:56:04 crc kubenswrapper[4778]: I0312 13:56:04.379237 4778 generic.go:334] "Generic (PLEG): container finished" podID="90b32527-d7b2-4938-a8c2-882067947e78" containerID="f6e775ed356b4c920e47d4cd6b52c164df8562cf9b83a71ba23edcf8ae60ceb9" exitCode=0 Mar 12 13:56:04 crc kubenswrapper[4778]: I0312 13:56:04.379395 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" event={"ID":"90b32527-d7b2-4938-a8c2-882067947e78","Type":"ContainerDied","Data":"f6e775ed356b4c920e47d4cd6b52c164df8562cf9b83a71ba23edcf8ae60ceb9"} Mar 12 13:56:05 crc kubenswrapper[4778]: I0312 13:56:05.748528 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:05 crc kubenswrapper[4778]: I0312 13:56:05.938853 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59ctl\" (UniqueName: \"kubernetes.io/projected/90b32527-d7b2-4938-a8c2-882067947e78-kube-api-access-59ctl\") pod \"90b32527-d7b2-4938-a8c2-882067947e78\" (UID: \"90b32527-d7b2-4938-a8c2-882067947e78\") " Mar 12 13:56:05 crc kubenswrapper[4778]: I0312 13:56:05.948619 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90b32527-d7b2-4938-a8c2-882067947e78-kube-api-access-59ctl" (OuterVolumeSpecName: "kube-api-access-59ctl") pod "90b32527-d7b2-4938-a8c2-882067947e78" (UID: "90b32527-d7b2-4938-a8c2-882067947e78"). InnerVolumeSpecName "kube-api-access-59ctl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:56:06 crc kubenswrapper[4778]: I0312 13:56:06.041776 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59ctl\" (UniqueName: \"kubernetes.io/projected/90b32527-d7b2-4938-a8c2-882067947e78-kube-api-access-59ctl\") on node \"crc\" DevicePath \"\"" Mar 12 13:56:06 crc kubenswrapper[4778]: I0312 13:56:06.397860 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" event={"ID":"90b32527-d7b2-4938-a8c2-882067947e78","Type":"ContainerDied","Data":"4fba4a78ca2fd2f35531ebdaf442cdd6b4b4c347b21e38401f16971eae6edcc0"} Mar 12 13:56:06 crc kubenswrapper[4778]: I0312 13:56:06.397905 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555396-lhqkd" Mar 12 13:56:06 crc kubenswrapper[4778]: I0312 13:56:06.397911 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fba4a78ca2fd2f35531ebdaf442cdd6b4b4c347b21e38401f16971eae6edcc0" Mar 12 13:56:06 crc kubenswrapper[4778]: I0312 13:56:06.822362 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555390-dml9r"] Mar 12 13:56:06 crc kubenswrapper[4778]: I0312 13:56:06.830619 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555390-dml9r"] Mar 12 13:56:08 crc kubenswrapper[4778]: I0312 13:56:08.265115 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b911b49-7b0f-48ef-9626-cd43d308d596" path="/var/lib/kubelet/pods/8b911b49-7b0f-48ef-9626-cd43d308d596/volumes" Mar 12 13:56:19 crc kubenswrapper[4778]: I0312 13:56:19.960630 4778 scope.go:117] "RemoveContainer" containerID="63aaec5f1f507e8a81d7498ca66c0663cdb9bde37e98025da0e464b4ce6c885e" Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.558250 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.558596 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.558649 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.559445 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e20e6fa2d381e3ff917a0f6074e27521c909a7932045eacfc15c005ed843cb93"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.559499 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://e20e6fa2d381e3ff917a0f6074e27521c909a7932045eacfc15c005ed843cb93" gracePeriod=600 Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.696158 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="e20e6fa2d381e3ff917a0f6074e27521c909a7932045eacfc15c005ed843cb93" exitCode=0 Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.696223 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"e20e6fa2d381e3ff917a0f6074e27521c909a7932045eacfc15c005ed843cb93"} Mar 12 13:56:28 crc kubenswrapper[4778]: I0312 13:56:28.696281 4778 scope.go:117] "RemoveContainer" containerID="5d7d3c0b73016a8d7ee117c8146ea559fc88bdaa58f9d10b5498b859a6d9fa8f" Mar 12 13:56:29 crc kubenswrapper[4778]: I0312 13:56:29.705664 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0"} Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.643155 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k8np5"] Mar 12 13:56:45 crc kubenswrapper[4778]: E0312 13:56:45.645619 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90b32527-d7b2-4938-a8c2-882067947e78" containerName="oc" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.645728 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="90b32527-d7b2-4938-a8c2-882067947e78" containerName="oc" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.646070 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="90b32527-d7b2-4938-a8c2-882067947e78" containerName="oc" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.647903 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.808762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2dxz\" (UniqueName: \"kubernetes.io/projected/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-kube-api-access-s2dxz\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.809086 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-catalog-content\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.809145 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-utilities\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.818066 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8np5"] Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.912022 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-catalog-content\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.912268 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-utilities\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.912402 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dxz\" (UniqueName: \"kubernetes.io/projected/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-kube-api-access-s2dxz\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.912939 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-catalog-content\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.912978 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-utilities\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.939613 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dxz\" (UniqueName: \"kubernetes.io/projected/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-kube-api-access-s2dxz\") pod \"redhat-operators-k8np5\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:45 crc kubenswrapper[4778]: I0312 13:56:45.972696 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:46 crc kubenswrapper[4778]: I0312 13:56:46.433020 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8np5"] Mar 12 13:56:46 crc kubenswrapper[4778]: I0312 13:56:46.853942 4778 generic.go:334] "Generic (PLEG): container finished" podID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerID="0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c" exitCode=0 Mar 12 13:56:46 crc kubenswrapper[4778]: I0312 13:56:46.854140 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerDied","Data":"0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c"} Mar 12 13:56:46 crc kubenswrapper[4778]: I0312 13:56:46.854306 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerStarted","Data":"9070894b7842f0aed54c5c06b02dc0798ae95d0cfa82fdac9ef69e082f746a53"} Mar 12 13:56:50 crc kubenswrapper[4778]: I0312 13:56:50.143153 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerStarted","Data":"4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703"} Mar 12 13:56:51 crc kubenswrapper[4778]: I0312 13:56:51.155053 4778 generic.go:334] "Generic (PLEG): container finished" podID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerID="4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703" exitCode=0 Mar 12 13:56:51 crc kubenswrapper[4778]: I0312 13:56:51.155097 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerDied","Data":"4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703"} Mar 12 13:56:53 crc kubenswrapper[4778]: I0312 13:56:53.178581 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerStarted","Data":"fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59"} Mar 12 13:56:55 crc kubenswrapper[4778]: I0312 13:56:55.974099 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:55 crc kubenswrapper[4778]: I0312 13:56:55.974640 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:56:57 crc kubenswrapper[4778]: I0312 13:56:57.022102 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k8np5" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="registry-server" probeResult="failure" output=< Mar 12 13:56:57 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 13:56:57 crc kubenswrapper[4778]: > Mar 12 13:57:06 crc kubenswrapper[4778]: I0312 13:57:06.014904 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:57:06 crc kubenswrapper[4778]: I0312 13:57:06.044051 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k8np5" podStartSLOduration=15.940376585 podStartE2EDuration="21.044031712s" podCreationTimestamp="2026-03-12 13:56:45 +0000 UTC" firstStartedPulling="2026-03-12 13:56:46.855710856 +0000 UTC m=+2825.304406252" lastFinishedPulling="2026-03-12 13:56:51.959365983 +0000 UTC m=+2830.408061379" observedRunningTime="2026-03-12 13:56:53.200219468 +0000 UTC m=+2831.648914864" watchObservedRunningTime="2026-03-12 13:57:06.044031712 +0000 UTC m=+2844.492727108" Mar 12 13:57:06 crc kubenswrapper[4778]: I0312 13:57:06.070115 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:57:06 crc kubenswrapper[4778]: I0312 13:57:06.249055 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8np5"] Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.294771 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k8np5" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="registry-server" containerID="cri-o://fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59" gracePeriod=2 Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.729483 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.780145 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-utilities\") pod \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.780396 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2dxz\" (UniqueName: \"kubernetes.io/projected/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-kube-api-access-s2dxz\") pod \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.780455 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-catalog-content\") pod \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\" (UID: \"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1\") " Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.782240 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-utilities" (OuterVolumeSpecName: "utilities") pod "1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" (UID: "1f9f45ee-d6ff-4369-b71a-1af75cc31ca1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.787850 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-kube-api-access-s2dxz" (OuterVolumeSpecName: "kube-api-access-s2dxz") pod "1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" (UID: "1f9f45ee-d6ff-4369-b71a-1af75cc31ca1"). InnerVolumeSpecName "kube-api-access-s2dxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.882704 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.882735 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2dxz\" (UniqueName: \"kubernetes.io/projected/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-kube-api-access-s2dxz\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.923102 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" (UID: "1f9f45ee-d6ff-4369-b71a-1af75cc31ca1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:57:07 crc kubenswrapper[4778]: I0312 13:57:07.985570 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.304135 4778 generic.go:334] "Generic (PLEG): container finished" podID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerID="fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59" exitCode=0 Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.304189 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8np5" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.304209 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerDied","Data":"fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59"} Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.304606 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8np5" event={"ID":"1f9f45ee-d6ff-4369-b71a-1af75cc31ca1","Type":"ContainerDied","Data":"9070894b7842f0aed54c5c06b02dc0798ae95d0cfa82fdac9ef69e082f746a53"} Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.304625 4778 scope.go:117] "RemoveContainer" containerID="fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.328132 4778 scope.go:117] "RemoveContainer" containerID="4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.335319 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8np5"] Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.345040 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k8np5"] Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.355706 4778 scope.go:117] "RemoveContainer" containerID="0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.388089 4778 scope.go:117] "RemoveContainer" containerID="fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59" Mar 12 13:57:08 crc kubenswrapper[4778]: E0312 13:57:08.388571 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59\": container with ID starting with fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59 not found: ID does not exist" containerID="fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.388620 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59"} err="failed to get container status \"fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59\": rpc error: code = NotFound desc = could not find container \"fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59\": container with ID starting with fe9eab63bd54027dc9747d671158f830246a8cbbcc67a324e1785af946b97a59 not found: ID does not exist" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.388648 4778 scope.go:117] "RemoveContainer" containerID="4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703" Mar 12 13:57:08 crc kubenswrapper[4778]: E0312 13:57:08.389070 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703\": container with ID starting with 4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703 not found: ID does not exist" containerID="4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.389214 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703"} err="failed to get container status \"4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703\": rpc error: code = NotFound desc = could not find container \"4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703\": container with ID starting with 4439d8ae883e35b8c6e7d2722cd2f00a49b0c79d4630243975e17bc91e0d2703 not found: ID does not exist" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.389312 4778 scope.go:117] "RemoveContainer" containerID="0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c" Mar 12 13:57:08 crc kubenswrapper[4778]: E0312 13:57:08.389697 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c\": container with ID starting with 0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c not found: ID does not exist" containerID="0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c" Mar 12 13:57:08 crc kubenswrapper[4778]: I0312 13:57:08.389743 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c"} err="failed to get container status \"0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c\": rpc error: code = NotFound desc = could not find container \"0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c\": container with ID starting with 0d24b81e7152e22db15d85c0639ccdf7f6a4a5d3388d523cdaa16266ec57d80c not found: ID does not exist" Mar 12 13:57:10 crc kubenswrapper[4778]: I0312 13:57:10.264228 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" path="/var/lib/kubelet/pods/1f9f45ee-d6ff-4369-b71a-1af75cc31ca1/volumes" Mar 12 13:57:23 crc kubenswrapper[4778]: I0312 13:57:23.797045 4778 generic.go:334] "Generic (PLEG): container finished" podID="6ed77f87-e6b2-4c7a-8b0e-003106200dc8" containerID="e2a35e751ce79cb5226fd46ca73472f5cd7c47201c7bf749d5ffd3dae25fcc72" exitCode=0 Mar 12 13:57:23 crc kubenswrapper[4778]: I0312 13:57:23.797122 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" event={"ID":"6ed77f87-e6b2-4c7a-8b0e-003106200dc8","Type":"ContainerDied","Data":"e2a35e751ce79cb5226fd46ca73472f5cd7c47201c7bf749d5ffd3dae25fcc72"} Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.282610 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.457988 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-1\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458085 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-combined-ca-bundle\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458116 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp6ls\" (UniqueName: \"kubernetes.io/projected/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-kube-api-access-fp6ls\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458142 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-inventory\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458267 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-3\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458305 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-0\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458359 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-0\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458384 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-2\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458434 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-ssh-key-openstack-edpm-ipam\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458464 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-1\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.458537 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-extra-config-0\") pod \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\" (UID: \"6ed77f87-e6b2-4c7a-8b0e-003106200dc8\") " Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.478593 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.478889 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-kube-api-access-fp6ls" (OuterVolumeSpecName: "kube-api-access-fp6ls") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "kube-api-access-fp6ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.491064 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.491152 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.491406 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.492313 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.493383 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.499958 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.502519 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.506538 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-inventory" (OuterVolumeSpecName: "inventory") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.510696 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "6ed77f87-e6b2-4c7a-8b0e-003106200dc8" (UID: "6ed77f87-e6b2-4c7a-8b0e-003106200dc8"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560325 4778 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560548 4778 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560700 4778 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560766 4778 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560824 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560884 4778 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560943 4778 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.560999 4778 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.561058 4778 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.561110 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp6ls\" (UniqueName: \"kubernetes.io/projected/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-kube-api-access-fp6ls\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.561169 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ed77f87-e6b2-4c7a-8b0e-003106200dc8-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.822108 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" event={"ID":"6ed77f87-e6b2-4c7a-8b0e-003106200dc8","Type":"ContainerDied","Data":"1617b96c98df28869b5a069f5b74bb8126ce4a98898565a1251dadc01020d162"} Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.822460 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1617b96c98df28869b5a069f5b74bb8126ce4a98898565a1251dadc01020d162" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.822173 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5tw6s" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985129 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s"] Mar 12 13:57:25 crc kubenswrapper[4778]: E0312 13:57:25.985587 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="registry-server" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985610 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="registry-server" Mar 12 13:57:25 crc kubenswrapper[4778]: E0312 13:57:25.985629 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed77f87-e6b2-4c7a-8b0e-003106200dc8" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985641 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed77f87-e6b2-4c7a-8b0e-003106200dc8" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 12 13:57:25 crc kubenswrapper[4778]: E0312 13:57:25.985660 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="extract-utilities" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985669 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="extract-utilities" Mar 12 13:57:25 crc kubenswrapper[4778]: E0312 13:57:25.985694 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="extract-content" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985701 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="extract-content" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985928 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9f45ee-d6ff-4369-b71a-1af75cc31ca1" containerName="registry-server" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.985954 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed77f87-e6b2-4c7a-8b0e-003106200dc8" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.986755 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.991779 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.991805 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.992365 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.992563 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.993997 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qn2vx" Mar 12 13:57:25 crc kubenswrapper[4778]: I0312 13:57:25.997629 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s"] Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.171812 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.171897 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szp5b\" (UniqueName: \"kubernetes.io/projected/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-kube-api-access-szp5b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.171976 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.172028 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.172066 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.172136 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.172168 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.275355 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.275489 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.275740 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.276542 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.276588 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.276776 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.276829 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szp5b\" (UniqueName: \"kubernetes.io/projected/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-kube-api-access-szp5b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.281396 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.281397 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.281790 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.281898 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.282732 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.282860 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.298334 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szp5b\" (UniqueName: \"kubernetes.io/projected/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-kube-api-access-szp5b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.323835 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 13:57:26 crc kubenswrapper[4778]: I0312 13:57:26.847446 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s"] Mar 12 13:57:26 crc kubenswrapper[4778]: W0312 13:57:26.852800 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bfaafaf_36fb_4f1a_99ed_abb8b7bb4ae1.slice/crio-18c5ac6148c2a101f7ddccfc3e584f782b4abb6f5ed1d881a005c0fa17c5788f WatchSource:0}: Error finding container 18c5ac6148c2a101f7ddccfc3e584f782b4abb6f5ed1d881a005c0fa17c5788f: Status 404 returned error can't find the container with id 18c5ac6148c2a101f7ddccfc3e584f782b4abb6f5ed1d881a005c0fa17c5788f Mar 12 13:57:27 crc kubenswrapper[4778]: I0312 13:57:27.841161 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" event={"ID":"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1","Type":"ContainerStarted","Data":"a819986e64064460805e2e891a9205102c157b3069f167134a0bc6192d083ab6"} Mar 12 13:57:27 crc kubenswrapper[4778]: I0312 13:57:27.841216 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" event={"ID":"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1","Type":"ContainerStarted","Data":"18c5ac6148c2a101f7ddccfc3e584f782b4abb6f5ed1d881a005c0fa17c5788f"} Mar 12 13:57:27 crc kubenswrapper[4778]: I0312 13:57:27.871231 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" podStartSLOduration=2.433829751 podStartE2EDuration="2.871211205s" podCreationTimestamp="2026-03-12 13:57:25 +0000 UTC" firstStartedPulling="2026-03-12 13:57:26.860853045 +0000 UTC m=+2865.309548441" lastFinishedPulling="2026-03-12 13:57:27.298234509 +0000 UTC m=+2865.746929895" observedRunningTime="2026-03-12 13:57:27.869590648 +0000 UTC m=+2866.318286054" watchObservedRunningTime="2026-03-12 13:57:27.871211205 +0000 UTC m=+2866.319906601" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.146788 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555398-hhchd"] Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.148716 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.152935 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.153404 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.153697 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.160927 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555398-hhchd"] Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.223120 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22x86\" (UniqueName: \"kubernetes.io/projected/6a479324-f9a1-4095-a0b1-7c22fc72eb61-kube-api-access-22x86\") pod \"auto-csr-approver-29555398-hhchd\" (UID: \"6a479324-f9a1-4095-a0b1-7c22fc72eb61\") " pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.325657 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22x86\" (UniqueName: \"kubernetes.io/projected/6a479324-f9a1-4095-a0b1-7c22fc72eb61-kube-api-access-22x86\") pod \"auto-csr-approver-29555398-hhchd\" (UID: \"6a479324-f9a1-4095-a0b1-7c22fc72eb61\") " pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.343817 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22x86\" (UniqueName: \"kubernetes.io/projected/6a479324-f9a1-4095-a0b1-7c22fc72eb61-kube-api-access-22x86\") pod \"auto-csr-approver-29555398-hhchd\" (UID: \"6a479324-f9a1-4095-a0b1-7c22fc72eb61\") " pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.469961 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:00 crc kubenswrapper[4778]: I0312 13:58:00.967688 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555398-hhchd"] Mar 12 13:58:00 crc kubenswrapper[4778]: W0312 13:58:00.968723 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a479324_f9a1_4095_a0b1_7c22fc72eb61.slice/crio-f6c428e58cccf2f5ed6f1ed21daee2d015ece5d13f0cd9940356894ca7670473 WatchSource:0}: Error finding container f6c428e58cccf2f5ed6f1ed21daee2d015ece5d13f0cd9940356894ca7670473: Status 404 returned error can't find the container with id f6c428e58cccf2f5ed6f1ed21daee2d015ece5d13f0cd9940356894ca7670473 Mar 12 13:58:01 crc kubenswrapper[4778]: I0312 13:58:01.123279 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555398-hhchd" event={"ID":"6a479324-f9a1-4095-a0b1-7c22fc72eb61","Type":"ContainerStarted","Data":"f6c428e58cccf2f5ed6f1ed21daee2d015ece5d13f0cd9940356894ca7670473"} Mar 12 13:58:03 crc kubenswrapper[4778]: I0312 13:58:03.157887 4778 generic.go:334] "Generic (PLEG): container finished" podID="6a479324-f9a1-4095-a0b1-7c22fc72eb61" containerID="7f07e770195234611f35ac5fc4d8c046a4e07dd2e554f881a1c216e51689e210" exitCode=0 Mar 12 13:58:03 crc kubenswrapper[4778]: I0312 13:58:03.158013 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555398-hhchd" event={"ID":"6a479324-f9a1-4095-a0b1-7c22fc72eb61","Type":"ContainerDied","Data":"7f07e770195234611f35ac5fc4d8c046a4e07dd2e554f881a1c216e51689e210"} Mar 12 13:58:04 crc kubenswrapper[4778]: I0312 13:58:04.638640 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:04 crc kubenswrapper[4778]: I0312 13:58:04.726020 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22x86\" (UniqueName: \"kubernetes.io/projected/6a479324-f9a1-4095-a0b1-7c22fc72eb61-kube-api-access-22x86\") pod \"6a479324-f9a1-4095-a0b1-7c22fc72eb61\" (UID: \"6a479324-f9a1-4095-a0b1-7c22fc72eb61\") " Mar 12 13:58:04 crc kubenswrapper[4778]: I0312 13:58:04.732140 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a479324-f9a1-4095-a0b1-7c22fc72eb61-kube-api-access-22x86" (OuterVolumeSpecName: "kube-api-access-22x86") pod "6a479324-f9a1-4095-a0b1-7c22fc72eb61" (UID: "6a479324-f9a1-4095-a0b1-7c22fc72eb61"). InnerVolumeSpecName "kube-api-access-22x86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:58:04 crc kubenswrapper[4778]: I0312 13:58:04.828427 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22x86\" (UniqueName: \"kubernetes.io/projected/6a479324-f9a1-4095-a0b1-7c22fc72eb61-kube-api-access-22x86\") on node \"crc\" DevicePath \"\"" Mar 12 13:58:05 crc kubenswrapper[4778]: I0312 13:58:05.180342 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555398-hhchd" event={"ID":"6a479324-f9a1-4095-a0b1-7c22fc72eb61","Type":"ContainerDied","Data":"f6c428e58cccf2f5ed6f1ed21daee2d015ece5d13f0cd9940356894ca7670473"} Mar 12 13:58:05 crc kubenswrapper[4778]: I0312 13:58:05.180401 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6c428e58cccf2f5ed6f1ed21daee2d015ece5d13f0cd9940356894ca7670473" Mar 12 13:58:05 crc kubenswrapper[4778]: I0312 13:58:05.180438 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555398-hhchd" Mar 12 13:58:05 crc kubenswrapper[4778]: I0312 13:58:05.711756 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555392-wg78w"] Mar 12 13:58:05 crc kubenswrapper[4778]: I0312 13:58:05.723627 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555392-wg78w"] Mar 12 13:58:06 crc kubenswrapper[4778]: I0312 13:58:06.266832 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a0b5070-03d8-45fe-8148-c39a9b560fbb" path="/var/lib/kubelet/pods/0a0b5070-03d8-45fe-8148-c39a9b560fbb/volumes" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.155741 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4gq97"] Mar 12 13:58:13 crc kubenswrapper[4778]: E0312 13:58:13.156912 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a479324-f9a1-4095-a0b1-7c22fc72eb61" containerName="oc" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.156930 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a479324-f9a1-4095-a0b1-7c22fc72eb61" containerName="oc" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.157282 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a479324-f9a1-4095-a0b1-7c22fc72eb61" containerName="oc" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.159338 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.167949 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gq97"] Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.286368 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-catalog-content\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.286419 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-utilities\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.286443 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc9p5\" (UniqueName: \"kubernetes.io/projected/aa7fbd44-1786-4285-a062-f10d0971645b-kube-api-access-hc9p5\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.388756 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-catalog-content\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.388816 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-utilities\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.388842 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc9p5\" (UniqueName: \"kubernetes.io/projected/aa7fbd44-1786-4285-a062-f10d0971645b-kube-api-access-hc9p5\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.389660 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-catalog-content\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.389721 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-utilities\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.412624 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc9p5\" (UniqueName: \"kubernetes.io/projected/aa7fbd44-1786-4285-a062-f10d0971645b-kube-api-access-hc9p5\") pod \"redhat-marketplace-4gq97\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.480514 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:13 crc kubenswrapper[4778]: I0312 13:58:13.954693 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gq97"] Mar 12 13:58:14 crc kubenswrapper[4778]: I0312 13:58:14.266600 4778 generic.go:334] "Generic (PLEG): container finished" podID="aa7fbd44-1786-4285-a062-f10d0971645b" containerID="c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408" exitCode=0 Mar 12 13:58:14 crc kubenswrapper[4778]: I0312 13:58:14.266642 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerDied","Data":"c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408"} Mar 12 13:58:14 crc kubenswrapper[4778]: I0312 13:58:14.266666 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerStarted","Data":"e99743205e6d113d16d820625aa3ddcd4fd4ebfa7d55f88ae708417e22c46325"} Mar 12 13:58:15 crc kubenswrapper[4778]: I0312 13:58:15.284745 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerStarted","Data":"f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823"} Mar 12 13:58:16 crc kubenswrapper[4778]: I0312 13:58:16.297775 4778 generic.go:334] "Generic (PLEG): container finished" podID="aa7fbd44-1786-4285-a062-f10d0971645b" containerID="f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823" exitCode=0 Mar 12 13:58:16 crc kubenswrapper[4778]: I0312 13:58:16.297864 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerDied","Data":"f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823"} Mar 12 13:58:17 crc kubenswrapper[4778]: I0312 13:58:17.309817 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerStarted","Data":"d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56"} Mar 12 13:58:17 crc kubenswrapper[4778]: I0312 13:58:17.332395 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4gq97" podStartSLOduration=1.8536014889999999 podStartE2EDuration="4.332377105s" podCreationTimestamp="2026-03-12 13:58:13 +0000 UTC" firstStartedPulling="2026-03-12 13:58:14.268622413 +0000 UTC m=+2912.717317809" lastFinishedPulling="2026-03-12 13:58:16.747398029 +0000 UTC m=+2915.196093425" observedRunningTime="2026-03-12 13:58:17.330895923 +0000 UTC m=+2915.779591319" watchObservedRunningTime="2026-03-12 13:58:17.332377105 +0000 UTC m=+2915.781072511" Mar 12 13:58:20 crc kubenswrapper[4778]: I0312 13:58:20.086914 4778 scope.go:117] "RemoveContainer" containerID="1f9b06fe647c9c9d52674fc3e58e1c9d5c930036da2b4f235a350fc83217496f" Mar 12 13:58:23 crc kubenswrapper[4778]: I0312 13:58:23.481138 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:23 crc kubenswrapper[4778]: I0312 13:58:23.482437 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:23 crc kubenswrapper[4778]: I0312 13:58:23.528450 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:24 crc kubenswrapper[4778]: I0312 13:58:24.424127 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:24 crc kubenswrapper[4778]: I0312 13:58:24.475838 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gq97"] Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.388876 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4gq97" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="registry-server" containerID="cri-o://d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56" gracePeriod=2 Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.881924 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.924523 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc9p5\" (UniqueName: \"kubernetes.io/projected/aa7fbd44-1786-4285-a062-f10d0971645b-kube-api-access-hc9p5\") pod \"aa7fbd44-1786-4285-a062-f10d0971645b\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.924635 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-catalog-content\") pod \"aa7fbd44-1786-4285-a062-f10d0971645b\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.924688 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-utilities\") pod \"aa7fbd44-1786-4285-a062-f10d0971645b\" (UID: \"aa7fbd44-1786-4285-a062-f10d0971645b\") " Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.925730 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-utilities" (OuterVolumeSpecName: "utilities") pod "aa7fbd44-1786-4285-a062-f10d0971645b" (UID: "aa7fbd44-1786-4285-a062-f10d0971645b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.935042 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7fbd44-1786-4285-a062-f10d0971645b-kube-api-access-hc9p5" (OuterVolumeSpecName: "kube-api-access-hc9p5") pod "aa7fbd44-1786-4285-a062-f10d0971645b" (UID: "aa7fbd44-1786-4285-a062-f10d0971645b"). InnerVolumeSpecName "kube-api-access-hc9p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 13:58:26 crc kubenswrapper[4778]: I0312 13:58:26.956711 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa7fbd44-1786-4285-a062-f10d0971645b" (UID: "aa7fbd44-1786-4285-a062-f10d0971645b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.027371 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.027416 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc9p5\" (UniqueName: \"kubernetes.io/projected/aa7fbd44-1786-4285-a062-f10d0971645b-kube-api-access-hc9p5\") on node \"crc\" DevicePath \"\"" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.027434 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7fbd44-1786-4285-a062-f10d0971645b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.398985 4778 generic.go:334] "Generic (PLEG): container finished" podID="aa7fbd44-1786-4285-a062-f10d0971645b" containerID="d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56" exitCode=0 Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.399026 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerDied","Data":"d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56"} Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.399030 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gq97" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.399058 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gq97" event={"ID":"aa7fbd44-1786-4285-a062-f10d0971645b","Type":"ContainerDied","Data":"e99743205e6d113d16d820625aa3ddcd4fd4ebfa7d55f88ae708417e22c46325"} Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.399074 4778 scope.go:117] "RemoveContainer" containerID="d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.421038 4778 scope.go:117] "RemoveContainer" containerID="f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.434588 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gq97"] Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.446131 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gq97"] Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.454935 4778 scope.go:117] "RemoveContainer" containerID="c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.489653 4778 scope.go:117] "RemoveContainer" containerID="d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56" Mar 12 13:58:27 crc kubenswrapper[4778]: E0312 13:58:27.490153 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56\": container with ID starting with d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56 not found: ID does not exist" containerID="d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.490279 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56"} err="failed to get container status \"d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56\": rpc error: code = NotFound desc = could not find container \"d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56\": container with ID starting with d51ebac645c914d827233a7e05640db3535b9bea5531b876c39672135af80a56 not found: ID does not exist" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.490304 4778 scope.go:117] "RemoveContainer" containerID="f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823" Mar 12 13:58:27 crc kubenswrapper[4778]: E0312 13:58:27.490577 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823\": container with ID starting with f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823 not found: ID does not exist" containerID="f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.490611 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823"} err="failed to get container status \"f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823\": rpc error: code = NotFound desc = could not find container \"f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823\": container with ID starting with f3fddcc47109eae320c7476971617745b962d7a8925cfec715a7e6160df6a823 not found: ID does not exist" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.490630 4778 scope.go:117] "RemoveContainer" containerID="c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408" Mar 12 13:58:27 crc kubenswrapper[4778]: E0312 13:58:27.490996 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408\": container with ID starting with c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408 not found: ID does not exist" containerID="c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408" Mar 12 13:58:27 crc kubenswrapper[4778]: I0312 13:58:27.491021 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408"} err="failed to get container status \"c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408\": rpc error: code = NotFound desc = could not find container \"c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408\": container with ID starting with c586eba3448e91b65e6147aa9ecb68c6b2bc24f9a1b32cc0aebf52545c4c6408 not found: ID does not exist" Mar 12 13:58:28 crc kubenswrapper[4778]: I0312 13:58:28.263817 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" path="/var/lib/kubelet/pods/aa7fbd44-1786-4285-a062-f10d0971645b/volumes" Mar 12 13:58:28 crc kubenswrapper[4778]: I0312 13:58:28.558007 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:58:28 crc kubenswrapper[4778]: I0312 13:58:28.558065 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:58:58 crc kubenswrapper[4778]: I0312 13:58:58.558451 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:58:58 crc kubenswrapper[4778]: I0312 13:58:58.559084 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:59:28 crc kubenswrapper[4778]: I0312 13:59:28.557687 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 13:59:28 crc kubenswrapper[4778]: I0312 13:59:28.558237 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 13:59:28 crc kubenswrapper[4778]: I0312 13:59:28.558290 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 13:59:28 crc kubenswrapper[4778]: I0312 13:59:28.559238 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 13:59:28 crc kubenswrapper[4778]: I0312 13:59:28.559289 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" gracePeriod=600 Mar 12 13:59:28 crc kubenswrapper[4778]: E0312 13:59:28.687979 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:59:29 crc kubenswrapper[4778]: I0312 13:59:29.276491 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" exitCode=0 Mar 12 13:59:29 crc kubenswrapper[4778]: I0312 13:59:29.276524 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0"} Mar 12 13:59:29 crc kubenswrapper[4778]: I0312 13:59:29.276576 4778 scope.go:117] "RemoveContainer" containerID="e20e6fa2d381e3ff917a0f6074e27521c909a7932045eacfc15c005ed843cb93" Mar 12 13:59:29 crc kubenswrapper[4778]: I0312 13:59:29.277335 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 13:59:29 crc kubenswrapper[4778]: E0312 13:59:29.277646 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:59:43 crc kubenswrapper[4778]: I0312 13:59:43.308306 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 13:59:43 crc kubenswrapper[4778]: E0312 13:59:43.309040 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 13:59:58 crc kubenswrapper[4778]: I0312 13:59:58.254483 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 13:59:58 crc kubenswrapper[4778]: E0312 13:59:58.255515 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.156601 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555400-c5pzt"] Mar 12 14:00:00 crc kubenswrapper[4778]: E0312 14:00:00.158925 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="extract-content" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.158958 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="extract-content" Mar 12 14:00:00 crc kubenswrapper[4778]: E0312 14:00:00.158974 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="extract-utilities" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.158983 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="extract-utilities" Mar 12 14:00:00 crc kubenswrapper[4778]: E0312 14:00:00.159019 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="registry-server" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.159030 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="registry-server" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.159312 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7fbd44-1786-4285-a062-f10d0971645b" containerName="registry-server" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.160276 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.280389 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.280726 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.293921 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.297842 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555400-c5pzt"] Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.323882 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8"] Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.325435 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.327733 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.328233 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.340335 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt5fv\" (UniqueName: \"kubernetes.io/projected/2425d74f-ef53-43bc-8c8f-976333a9cc6a-kube-api-access-kt5fv\") pod \"auto-csr-approver-29555400-c5pzt\" (UID: \"2425d74f-ef53-43bc-8c8f-976333a9cc6a\") " pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.345518 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8"] Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.442955 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt5fv\" (UniqueName: \"kubernetes.io/projected/2425d74f-ef53-43bc-8c8f-976333a9cc6a-kube-api-access-kt5fv\") pod \"auto-csr-approver-29555400-c5pzt\" (UID: \"2425d74f-ef53-43bc-8c8f-976333a9cc6a\") " pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.443036 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jkzm\" (UniqueName: \"kubernetes.io/projected/5d85560c-89e4-4723-beb0-aeda87d0791a-kube-api-access-6jkzm\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.443093 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d85560c-89e4-4723-beb0-aeda87d0791a-secret-volume\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.443142 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d85560c-89e4-4723-beb0-aeda87d0791a-config-volume\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.465027 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt5fv\" (UniqueName: \"kubernetes.io/projected/2425d74f-ef53-43bc-8c8f-976333a9cc6a-kube-api-access-kt5fv\") pod \"auto-csr-approver-29555400-c5pzt\" (UID: \"2425d74f-ef53-43bc-8c8f-976333a9cc6a\") " pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.544534 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d85560c-89e4-4723-beb0-aeda87d0791a-config-volume\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.545729 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jkzm\" (UniqueName: \"kubernetes.io/projected/5d85560c-89e4-4723-beb0-aeda87d0791a-kube-api-access-6jkzm\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.545842 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d85560c-89e4-4723-beb0-aeda87d0791a-secret-volume\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.546369 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d85560c-89e4-4723-beb0-aeda87d0791a-config-volume\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.559459 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d85560c-89e4-4723-beb0-aeda87d0791a-secret-volume\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.563623 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jkzm\" (UniqueName: \"kubernetes.io/projected/5d85560c-89e4-4723-beb0-aeda87d0791a-kube-api-access-6jkzm\") pod \"collect-profiles-29555400-lrxd8\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.623168 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:00 crc kubenswrapper[4778]: I0312 14:00:00.645743 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:01 crc kubenswrapper[4778]: I0312 14:00:01.090136 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555400-c5pzt"] Mar 12 14:00:01 crc kubenswrapper[4778]: I0312 14:00:01.147971 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8"] Mar 12 14:00:01 crc kubenswrapper[4778]: W0312 14:00:01.152014 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d85560c_89e4_4723_beb0_aeda87d0791a.slice/crio-9df22d6360280afbb45c874c3f1c87b75dd48d3df4b442954817d575f3b7402f WatchSource:0}: Error finding container 9df22d6360280afbb45c874c3f1c87b75dd48d3df4b442954817d575f3b7402f: Status 404 returned error can't find the container with id 9df22d6360280afbb45c874c3f1c87b75dd48d3df4b442954817d575f3b7402f Mar 12 14:00:01 crc kubenswrapper[4778]: I0312 14:00:01.601637 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" event={"ID":"2425d74f-ef53-43bc-8c8f-976333a9cc6a","Type":"ContainerStarted","Data":"09af7cee21adeff6c51ad2ef43b20a2fcd251532a1429e097c4b6c46a2a78a68"} Mar 12 14:00:01 crc kubenswrapper[4778]: I0312 14:00:01.622602 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" event={"ID":"5d85560c-89e4-4723-beb0-aeda87d0791a","Type":"ContainerStarted","Data":"96aa4949ff208afd2c193ba8303ee15ee08731bdac5eecb0faaa4ff029a2c93a"} Mar 12 14:00:01 crc kubenswrapper[4778]: I0312 14:00:01.622672 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" event={"ID":"5d85560c-89e4-4723-beb0-aeda87d0791a","Type":"ContainerStarted","Data":"9df22d6360280afbb45c874c3f1c87b75dd48d3df4b442954817d575f3b7402f"} Mar 12 14:00:01 crc kubenswrapper[4778]: I0312 14:00:01.649314 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" podStartSLOduration=1.649294212 podStartE2EDuration="1.649294212s" podCreationTimestamp="2026-03-12 14:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 14:00:01.645968618 +0000 UTC m=+3020.094664014" watchObservedRunningTime="2026-03-12 14:00:01.649294212 +0000 UTC m=+3020.097989608" Mar 12 14:00:02 crc kubenswrapper[4778]: I0312 14:00:02.632676 4778 generic.go:334] "Generic (PLEG): container finished" podID="5d85560c-89e4-4723-beb0-aeda87d0791a" containerID="96aa4949ff208afd2c193ba8303ee15ee08731bdac5eecb0faaa4ff029a2c93a" exitCode=0 Mar 12 14:00:02 crc kubenswrapper[4778]: I0312 14:00:02.632741 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" event={"ID":"5d85560c-89e4-4723-beb0-aeda87d0791a","Type":"ContainerDied","Data":"96aa4949ff208afd2c193ba8303ee15ee08731bdac5eecb0faaa4ff029a2c93a"} Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.040200 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.323196 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d85560c-89e4-4723-beb0-aeda87d0791a-config-volume\") pod \"5d85560c-89e4-4723-beb0-aeda87d0791a\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.323306 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d85560c-89e4-4723-beb0-aeda87d0791a-secret-volume\") pod \"5d85560c-89e4-4723-beb0-aeda87d0791a\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.323386 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jkzm\" (UniqueName: \"kubernetes.io/projected/5d85560c-89e4-4723-beb0-aeda87d0791a-kube-api-access-6jkzm\") pod \"5d85560c-89e4-4723-beb0-aeda87d0791a\" (UID: \"5d85560c-89e4-4723-beb0-aeda87d0791a\") " Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.335198 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d85560c-89e4-4723-beb0-aeda87d0791a-config-volume" (OuterVolumeSpecName: "config-volume") pod "5d85560c-89e4-4723-beb0-aeda87d0791a" (UID: "5d85560c-89e4-4723-beb0-aeda87d0791a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.342139 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d85560c-89e4-4723-beb0-aeda87d0791a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.378907 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d85560c-89e4-4723-beb0-aeda87d0791a-kube-api-access-6jkzm" (OuterVolumeSpecName: "kube-api-access-6jkzm") pod "5d85560c-89e4-4723-beb0-aeda87d0791a" (UID: "5d85560c-89e4-4723-beb0-aeda87d0791a"). InnerVolumeSpecName "kube-api-access-6jkzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.393361 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d85560c-89e4-4723-beb0-aeda87d0791a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5d85560c-89e4-4723-beb0-aeda87d0791a" (UID: "5d85560c-89e4-4723-beb0-aeda87d0791a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.451731 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d85560c-89e4-4723-beb0-aeda87d0791a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.451953 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jkzm\" (UniqueName: \"kubernetes.io/projected/5d85560c-89e4-4723-beb0-aeda87d0791a-kube-api-access-6jkzm\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.654860 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" event={"ID":"5d85560c-89e4-4723-beb0-aeda87d0791a","Type":"ContainerDied","Data":"9df22d6360280afbb45c874c3f1c87b75dd48d3df4b442954817d575f3b7402f"} Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.654910 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9df22d6360280afbb45c874c3f1c87b75dd48d3df4b442954817d575f3b7402f" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.654982 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8" Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.710732 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555355-68226"] Mar 12 14:00:04 crc kubenswrapper[4778]: I0312 14:00:04.718713 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555355-68226"] Mar 12 14:00:05 crc kubenswrapper[4778]: I0312 14:00:05.668626 4778 generic.go:334] "Generic (PLEG): container finished" podID="2425d74f-ef53-43bc-8c8f-976333a9cc6a" containerID="363f3ad00ca01b087e83fcbce9630716537dd1aa2dde624be9a2f51cfec1e8a6" exitCode=0 Mar 12 14:00:05 crc kubenswrapper[4778]: I0312 14:00:05.668708 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" event={"ID":"2425d74f-ef53-43bc-8c8f-976333a9cc6a","Type":"ContainerDied","Data":"363f3ad00ca01b087e83fcbce9630716537dd1aa2dde624be9a2f51cfec1e8a6"} Mar 12 14:00:06 crc kubenswrapper[4778]: I0312 14:00:06.265311 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6197b3a9-f02f-4e5d-8196-b617fffa467d" path="/var/lib/kubelet/pods/6197b3a9-f02f-4e5d-8196-b617fffa467d/volumes" Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.106573 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.268477 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt5fv\" (UniqueName: \"kubernetes.io/projected/2425d74f-ef53-43bc-8c8f-976333a9cc6a-kube-api-access-kt5fv\") pod \"2425d74f-ef53-43bc-8c8f-976333a9cc6a\" (UID: \"2425d74f-ef53-43bc-8c8f-976333a9cc6a\") " Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.278038 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2425d74f-ef53-43bc-8c8f-976333a9cc6a-kube-api-access-kt5fv" (OuterVolumeSpecName: "kube-api-access-kt5fv") pod "2425d74f-ef53-43bc-8c8f-976333a9cc6a" (UID: "2425d74f-ef53-43bc-8c8f-976333a9cc6a"). InnerVolumeSpecName "kube-api-access-kt5fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.372035 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt5fv\" (UniqueName: \"kubernetes.io/projected/2425d74f-ef53-43bc-8c8f-976333a9cc6a-kube-api-access-kt5fv\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.693227 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" event={"ID":"2425d74f-ef53-43bc-8c8f-976333a9cc6a","Type":"ContainerDied","Data":"09af7cee21adeff6c51ad2ef43b20a2fcd251532a1429e097c4b6c46a2a78a68"} Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.693453 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09af7cee21adeff6c51ad2ef43b20a2fcd251532a1429e097c4b6c46a2a78a68" Mar 12 14:00:07 crc kubenswrapper[4778]: I0312 14:00:07.693323 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555400-c5pzt" Mar 12 14:00:08 crc kubenswrapper[4778]: I0312 14:00:08.162099 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555394-7f7nf"] Mar 12 14:00:08 crc kubenswrapper[4778]: I0312 14:00:08.172716 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555394-7f7nf"] Mar 12 14:00:08 crc kubenswrapper[4778]: I0312 14:00:08.270392 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65fbb68a-57a2-40bf-9149-6cfe13fe147c" path="/var/lib/kubelet/pods/65fbb68a-57a2-40bf-9149-6cfe13fe147c/volumes" Mar 12 14:00:11 crc kubenswrapper[4778]: I0312 14:00:11.326143 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:00:11 crc kubenswrapper[4778]: E0312 14:00:11.326828 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:00:20 crc kubenswrapper[4778]: I0312 14:00:20.187934 4778 scope.go:117] "RemoveContainer" containerID="3954f4afdb430b04a44fc16681134a45669f465399452c67b26950fbb78cb40a" Mar 12 14:00:20 crc kubenswrapper[4778]: I0312 14:00:20.213278 4778 scope.go:117] "RemoveContainer" containerID="b93a8a130b5f9b7d0852157c6942677a4b8f445ae1cc7062b429977ab9491779" Mar 12 14:00:22 crc kubenswrapper[4778]: I0312 14:00:22.261819 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:00:22 crc kubenswrapper[4778]: E0312 14:00:22.262497 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:00:34 crc kubenswrapper[4778]: I0312 14:00:34.001767 4778 generic.go:334] "Generic (PLEG): container finished" podID="2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" containerID="a819986e64064460805e2e891a9205102c157b3069f167134a0bc6192d083ab6" exitCode=0 Mar 12 14:00:34 crc kubenswrapper[4778]: I0312 14:00:34.001834 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" event={"ID":"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1","Type":"ContainerDied","Data":"a819986e64064460805e2e891a9205102c157b3069f167134a0bc6192d083ab6"} Mar 12 14:00:34 crc kubenswrapper[4778]: I0312 14:00:34.254070 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:00:34 crc kubenswrapper[4778]: E0312 14:00:34.254337 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.400689 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468372 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-0\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468482 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-1\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468529 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-2\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468550 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-inventory\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468586 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ssh-key-openstack-edpm-ipam\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468633 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-telemetry-combined-ca-bundle\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.468724 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szp5b\" (UniqueName: \"kubernetes.io/projected/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-kube-api-access-szp5b\") pod \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\" (UID: \"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1\") " Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.474492 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.474923 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-kube-api-access-szp5b" (OuterVolumeSpecName: "kube-api-access-szp5b") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "kube-api-access-szp5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.496265 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.498301 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.500091 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.500510 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.501870 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-inventory" (OuterVolumeSpecName: "inventory") pod "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" (UID: "2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.570969 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szp5b\" (UniqueName: \"kubernetes.io/projected/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-kube-api-access-szp5b\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.571302 4778 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.571314 4778 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.571323 4778 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.571335 4778 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.571344 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:35 crc kubenswrapper[4778]: I0312 14:00:35.571354 4778 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 14:00:36 crc kubenswrapper[4778]: I0312 14:00:36.018544 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" event={"ID":"2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1","Type":"ContainerDied","Data":"18c5ac6148c2a101f7ddccfc3e584f782b4abb6f5ed1d881a005c0fa17c5788f"} Mar 12 14:00:36 crc kubenswrapper[4778]: I0312 14:00:36.018584 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18c5ac6148c2a101f7ddccfc3e584f782b4abb6f5ed1d881a005c0fa17c5788f" Mar 12 14:00:36 crc kubenswrapper[4778]: I0312 14:00:36.018611 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s" Mar 12 14:00:46 crc kubenswrapper[4778]: I0312 14:00:46.254252 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:00:46 crc kubenswrapper[4778]: E0312 14:00:46.255390 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:00:58 crc kubenswrapper[4778]: I0312 14:00:58.253987 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:00:58 crc kubenswrapper[4778]: E0312 14:00:58.254853 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.157620 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29555401-vjgkl"] Mar 12 14:01:00 crc kubenswrapper[4778]: E0312 14:01:00.158722 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.158743 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 12 14:01:00 crc kubenswrapper[4778]: E0312 14:01:00.158758 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d85560c-89e4-4723-beb0-aeda87d0791a" containerName="collect-profiles" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.158777 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d85560c-89e4-4723-beb0-aeda87d0791a" containerName="collect-profiles" Mar 12 14:01:00 crc kubenswrapper[4778]: E0312 14:01:00.158790 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2425d74f-ef53-43bc-8c8f-976333a9cc6a" containerName="oc" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.158797 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2425d74f-ef53-43bc-8c8f-976333a9cc6a" containerName="oc" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.159041 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.159056 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2425d74f-ef53-43bc-8c8f-976333a9cc6a" containerName="oc" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.159080 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d85560c-89e4-4723-beb0-aeda87d0791a" containerName="collect-profiles" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.159892 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.198501 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29555401-vjgkl"] Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.225015 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-config-data\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.225092 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb9xz\" (UniqueName: \"kubernetes.io/projected/e4df6927-3452-4b36-b59a-a1fdcd4272a4-kube-api-access-lb9xz\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.225122 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-fernet-keys\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.225235 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-combined-ca-bundle\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.328091 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-combined-ca-bundle\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.332336 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-config-data\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.340037 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb9xz\" (UniqueName: \"kubernetes.io/projected/e4df6927-3452-4b36-b59a-a1fdcd4272a4-kube-api-access-lb9xz\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.340137 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-fernet-keys\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.342541 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-combined-ca-bundle\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.343901 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-config-data\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.344467 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-fernet-keys\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.371213 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb9xz\" (UniqueName: \"kubernetes.io/projected/e4df6927-3452-4b36-b59a-a1fdcd4272a4-kube-api-access-lb9xz\") pod \"keystone-cron-29555401-vjgkl\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.513911 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:00 crc kubenswrapper[4778]: I0312 14:01:00.975969 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29555401-vjgkl"] Mar 12 14:01:01 crc kubenswrapper[4778]: I0312 14:01:01.236272 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555401-vjgkl" event={"ID":"e4df6927-3452-4b36-b59a-a1fdcd4272a4","Type":"ContainerStarted","Data":"9086b9928613830092038a1fc1873e5d9952f8b26aed48c1a8aeece2a4bbfb3a"} Mar 12 14:01:01 crc kubenswrapper[4778]: I0312 14:01:01.236318 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555401-vjgkl" event={"ID":"e4df6927-3452-4b36-b59a-a1fdcd4272a4","Type":"ContainerStarted","Data":"bf78dfb35e03fa1908b33d1f483749337571e17bce7c98498613e80b293582e2"} Mar 12 14:01:03 crc kubenswrapper[4778]: I0312 14:01:03.255980 4778 generic.go:334] "Generic (PLEG): container finished" podID="e4df6927-3452-4b36-b59a-a1fdcd4272a4" containerID="9086b9928613830092038a1fc1873e5d9952f8b26aed48c1a8aeece2a4bbfb3a" exitCode=0 Mar 12 14:01:03 crc kubenswrapper[4778]: I0312 14:01:03.256042 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555401-vjgkl" event={"ID":"e4df6927-3452-4b36-b59a-a1fdcd4272a4","Type":"ContainerDied","Data":"9086b9928613830092038a1fc1873e5d9952f8b26aed48c1a8aeece2a4bbfb3a"} Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.604822 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.627161 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-config-data\") pod \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.627281 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb9xz\" (UniqueName: \"kubernetes.io/projected/e4df6927-3452-4b36-b59a-a1fdcd4272a4-kube-api-access-lb9xz\") pod \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.627363 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-fernet-keys\") pod \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.627556 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-combined-ca-bundle\") pod \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\" (UID: \"e4df6927-3452-4b36-b59a-a1fdcd4272a4\") " Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.633394 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e4df6927-3452-4b36-b59a-a1fdcd4272a4" (UID: "e4df6927-3452-4b36-b59a-a1fdcd4272a4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.634041 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4df6927-3452-4b36-b59a-a1fdcd4272a4-kube-api-access-lb9xz" (OuterVolumeSpecName: "kube-api-access-lb9xz") pod "e4df6927-3452-4b36-b59a-a1fdcd4272a4" (UID: "e4df6927-3452-4b36-b59a-a1fdcd4272a4"). InnerVolumeSpecName "kube-api-access-lb9xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.663385 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4df6927-3452-4b36-b59a-a1fdcd4272a4" (UID: "e4df6927-3452-4b36-b59a-a1fdcd4272a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.688177 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-config-data" (OuterVolumeSpecName: "config-data") pod "e4df6927-3452-4b36-b59a-a1fdcd4272a4" (UID: "e4df6927-3452-4b36-b59a-a1fdcd4272a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.731335 4778 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.731373 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.731386 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4df6927-3452-4b36-b59a-a1fdcd4272a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 14:01:04 crc kubenswrapper[4778]: I0312 14:01:04.731397 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb9xz\" (UniqueName: \"kubernetes.io/projected/e4df6927-3452-4b36-b59a-a1fdcd4272a4-kube-api-access-lb9xz\") on node \"crc\" DevicePath \"\"" Mar 12 14:01:05 crc kubenswrapper[4778]: I0312 14:01:05.275140 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555401-vjgkl" event={"ID":"e4df6927-3452-4b36-b59a-a1fdcd4272a4","Type":"ContainerDied","Data":"bf78dfb35e03fa1908b33d1f483749337571e17bce7c98498613e80b293582e2"} Mar 12 14:01:05 crc kubenswrapper[4778]: I0312 14:01:05.275753 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf78dfb35e03fa1908b33d1f483749337571e17bce7c98498613e80b293582e2" Mar 12 14:01:05 crc kubenswrapper[4778]: I0312 14:01:05.275438 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555401-vjgkl" Mar 12 14:01:13 crc kubenswrapper[4778]: I0312 14:01:13.253963 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:01:13 crc kubenswrapper[4778]: E0312 14:01:13.254940 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:01:27 crc kubenswrapper[4778]: I0312 14:01:27.254681 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:01:27 crc kubenswrapper[4778]: E0312 14:01:27.255488 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.920335 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 12 14:01:33 crc kubenswrapper[4778]: E0312 14:01:33.921327 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4df6927-3452-4b36-b59a-a1fdcd4272a4" containerName="keystone-cron" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.921345 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4df6927-3452-4b36-b59a-a1fdcd4272a4" containerName="keystone-cron" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.921576 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4df6927-3452-4b36-b59a-a1fdcd4272a4" containerName="keystone-cron" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.922358 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.926060 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-s8dkq" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.926117 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.926218 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.938756 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 12 14:01:33 crc kubenswrapper[4778]: I0312 14:01:33.942167 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.012075 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.012211 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.012246 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.013205 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.013341 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.013378 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-config-data\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.013432 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ssdw\" (UniqueName: \"kubernetes.io/projected/74897d0a-ca7b-4589-bd4c-75910c2d491c-kube-api-access-4ssdw\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.013485 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.013668 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116012 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116073 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116132 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116165 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116280 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116320 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116343 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-config-data\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116383 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ssdw\" (UniqueName: \"kubernetes.io/projected/74897d0a-ca7b-4589-bd4c-75910c2d491c-kube-api-access-4ssdw\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116416 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.116900 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.117094 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.117769 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.117942 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.118066 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-config-data\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.136009 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.136353 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.143133 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.146712 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ssdw\" (UniqueName: \"kubernetes.io/projected/74897d0a-ca7b-4589-bd4c-75910c2d491c-kube-api-access-4ssdw\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.166800 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.257641 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.805741 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 12 14:01:34 crc kubenswrapper[4778]: I0312 14:01:34.813399 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:01:35 crc kubenswrapper[4778]: I0312 14:01:35.544887 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"74897d0a-ca7b-4589-bd4c-75910c2d491c","Type":"ContainerStarted","Data":"454ca901956127a4048551d166d33c00269e2d8a18f508b4b327654529c385c0"} Mar 12 14:01:42 crc kubenswrapper[4778]: I0312 14:01:42.271074 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:01:42 crc kubenswrapper[4778]: E0312 14:01:42.272452 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:01:53 crc kubenswrapper[4778]: I0312 14:01:53.254141 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:01:53 crc kubenswrapper[4778]: E0312 14:01:53.254688 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.159558 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555402-xtt9v"] Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.161671 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.164336 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.165318 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.166475 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.171549 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555402-xtt9v"] Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.298963 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gxb\" (UniqueName: \"kubernetes.io/projected/be917952-7177-4ef5-9efa-7858d1a11ded-kube-api-access-99gxb\") pod \"auto-csr-approver-29555402-xtt9v\" (UID: \"be917952-7177-4ef5-9efa-7858d1a11ded\") " pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.401026 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gxb\" (UniqueName: \"kubernetes.io/projected/be917952-7177-4ef5-9efa-7858d1a11ded-kube-api-access-99gxb\") pod \"auto-csr-approver-29555402-xtt9v\" (UID: \"be917952-7177-4ef5-9efa-7858d1a11ded\") " pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.428261 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gxb\" (UniqueName: \"kubernetes.io/projected/be917952-7177-4ef5-9efa-7858d1a11ded-kube-api-access-99gxb\") pod \"auto-csr-approver-29555402-xtt9v\" (UID: \"be917952-7177-4ef5-9efa-7858d1a11ded\") " pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:00 crc kubenswrapper[4778]: I0312 14:02:00.492418 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:04 crc kubenswrapper[4778]: E0312 14:02:04.854373 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 12 14:02:04 crc kubenswrapper[4778]: E0312 14:02:04.854991 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4ssdw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(74897d0a-ca7b-4589-bd4c-75910c2d491c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 14:02:04 crc kubenswrapper[4778]: E0312 14:02:04.858282 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="74897d0a-ca7b-4589-bd4c-75910c2d491c" Mar 12 14:02:05 crc kubenswrapper[4778]: I0312 14:02:05.208705 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555402-xtt9v"] Mar 12 14:02:05 crc kubenswrapper[4778]: I0312 14:02:05.834460 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" event={"ID":"be917952-7177-4ef5-9efa-7858d1a11ded","Type":"ContainerStarted","Data":"af14200afd084ab4265094e5fc43e707cf661ff1277fb7f0a1a629d498a2dca8"} Mar 12 14:02:05 crc kubenswrapper[4778]: E0312 14:02:05.836230 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="74897d0a-ca7b-4589-bd4c-75910c2d491c" Mar 12 14:02:06 crc kubenswrapper[4778]: I0312 14:02:06.254987 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:02:06 crc kubenswrapper[4778]: E0312 14:02:06.255584 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:02:07 crc kubenswrapper[4778]: I0312 14:02:07.856840 4778 generic.go:334] "Generic (PLEG): container finished" podID="be917952-7177-4ef5-9efa-7858d1a11ded" containerID="b626545edbe9764de0b916e68f0836b92c6dbff05d2ae4f9ae924f063217aca7" exitCode=0 Mar 12 14:02:07 crc kubenswrapper[4778]: I0312 14:02:07.856982 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" event={"ID":"be917952-7177-4ef5-9efa-7858d1a11ded","Type":"ContainerDied","Data":"b626545edbe9764de0b916e68f0836b92c6dbff05d2ae4f9ae924f063217aca7"} Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.189268 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.226595 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99gxb\" (UniqueName: \"kubernetes.io/projected/be917952-7177-4ef5-9efa-7858d1a11ded-kube-api-access-99gxb\") pod \"be917952-7177-4ef5-9efa-7858d1a11ded\" (UID: \"be917952-7177-4ef5-9efa-7858d1a11ded\") " Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.232884 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be917952-7177-4ef5-9efa-7858d1a11ded-kube-api-access-99gxb" (OuterVolumeSpecName: "kube-api-access-99gxb") pod "be917952-7177-4ef5-9efa-7858d1a11ded" (UID: "be917952-7177-4ef5-9efa-7858d1a11ded"). InnerVolumeSpecName "kube-api-access-99gxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.329201 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99gxb\" (UniqueName: \"kubernetes.io/projected/be917952-7177-4ef5-9efa-7858d1a11ded-kube-api-access-99gxb\") on node \"crc\" DevicePath \"\"" Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.873968 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" event={"ID":"be917952-7177-4ef5-9efa-7858d1a11ded","Type":"ContainerDied","Data":"af14200afd084ab4265094e5fc43e707cf661ff1277fb7f0a1a629d498a2dca8"} Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.874269 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af14200afd084ab4265094e5fc43e707cf661ff1277fb7f0a1a629d498a2dca8" Mar 12 14:02:09 crc kubenswrapper[4778]: I0312 14:02:09.874026 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555402-xtt9v" Mar 12 14:02:10 crc kubenswrapper[4778]: I0312 14:02:10.264715 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555396-lhqkd"] Mar 12 14:02:10 crc kubenswrapper[4778]: I0312 14:02:10.275575 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555396-lhqkd"] Mar 12 14:02:12 crc kubenswrapper[4778]: I0312 14:02:12.265811 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90b32527-d7b2-4938-a8c2-882067947e78" path="/var/lib/kubelet/pods/90b32527-d7b2-4938-a8c2-882067947e78/volumes" Mar 12 14:02:17 crc kubenswrapper[4778]: I0312 14:02:17.254514 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:02:17 crc kubenswrapper[4778]: E0312 14:02:17.255785 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:02:18 crc kubenswrapper[4778]: I0312 14:02:18.976201 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 12 14:02:20 crc kubenswrapper[4778]: I0312 14:02:20.350743 4778 scope.go:117] "RemoveContainer" containerID="f6e775ed356b4c920e47d4cd6b52c164df8562cf9b83a71ba23edcf8ae60ceb9" Mar 12 14:02:20 crc kubenswrapper[4778]: I0312 14:02:20.974486 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"74897d0a-ca7b-4589-bd4c-75910c2d491c","Type":"ContainerStarted","Data":"04824fe8df9ecfce713c8136bfb0516b3d49f4264b49ad91474ebd09ae740d91"} Mar 12 14:02:21 crc kubenswrapper[4778]: I0312 14:02:21.005330 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.845323112 podStartE2EDuration="49.005308291s" podCreationTimestamp="2026-03-12 14:01:32 +0000 UTC" firstStartedPulling="2026-03-12 14:01:34.813214932 +0000 UTC m=+3113.261910328" lastFinishedPulling="2026-03-12 14:02:18.973200101 +0000 UTC m=+3157.421895507" observedRunningTime="2026-03-12 14:02:20.995552244 +0000 UTC m=+3159.444247640" watchObservedRunningTime="2026-03-12 14:02:21.005308291 +0000 UTC m=+3159.454003687" Mar 12 14:02:31 crc kubenswrapper[4778]: I0312 14:02:31.253745 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:02:31 crc kubenswrapper[4778]: E0312 14:02:31.254625 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:02:46 crc kubenswrapper[4778]: I0312 14:02:46.254387 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:02:46 crc kubenswrapper[4778]: E0312 14:02:46.255135 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:03:00 crc kubenswrapper[4778]: I0312 14:03:00.253836 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:03:00 crc kubenswrapper[4778]: E0312 14:03:00.254684 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:03:13 crc kubenswrapper[4778]: I0312 14:03:13.253742 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:03:13 crc kubenswrapper[4778]: E0312 14:03:13.254528 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:03:25 crc kubenswrapper[4778]: I0312 14:03:25.255483 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:03:25 crc kubenswrapper[4778]: E0312 14:03:25.256428 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:03:40 crc kubenswrapper[4778]: I0312 14:03:40.254380 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:03:40 crc kubenswrapper[4778]: E0312 14:03:40.255079 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:03:55 crc kubenswrapper[4778]: I0312 14:03:55.254213 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:03:55 crc kubenswrapper[4778]: E0312 14:03:55.254882 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.176765 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555404-jwm56"] Mar 12 14:04:00 crc kubenswrapper[4778]: E0312 14:04:00.177933 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be917952-7177-4ef5-9efa-7858d1a11ded" containerName="oc" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.177950 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="be917952-7177-4ef5-9efa-7858d1a11ded" containerName="oc" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.178235 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="be917952-7177-4ef5-9efa-7858d1a11ded" containerName="oc" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.179095 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.181320 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.181653 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.182164 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.194345 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555404-jwm56"] Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.241386 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4zdm\" (UniqueName: \"kubernetes.io/projected/67604e51-359f-4c7f-b7df-a4f215a87085-kube-api-access-j4zdm\") pod \"auto-csr-approver-29555404-jwm56\" (UID: \"67604e51-359f-4c7f-b7df-a4f215a87085\") " pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.343894 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4zdm\" (UniqueName: \"kubernetes.io/projected/67604e51-359f-4c7f-b7df-a4f215a87085-kube-api-access-j4zdm\") pod \"auto-csr-approver-29555404-jwm56\" (UID: \"67604e51-359f-4c7f-b7df-a4f215a87085\") " pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.370474 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4zdm\" (UniqueName: \"kubernetes.io/projected/67604e51-359f-4c7f-b7df-a4f215a87085-kube-api-access-j4zdm\") pod \"auto-csr-approver-29555404-jwm56\" (UID: \"67604e51-359f-4c7f-b7df-a4f215a87085\") " pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:00 crc kubenswrapper[4778]: I0312 14:04:00.522937 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:01 crc kubenswrapper[4778]: I0312 14:04:01.004013 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555404-jwm56"] Mar 12 14:04:01 crc kubenswrapper[4778]: I0312 14:04:01.841563 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555404-jwm56" event={"ID":"67604e51-359f-4c7f-b7df-a4f215a87085","Type":"ContainerStarted","Data":"1b601cd7deaebbb8c485b6a4bc8cce371566ad54d4f0487a2fe0ff8eb085f00b"} Mar 12 14:04:02 crc kubenswrapper[4778]: I0312 14:04:02.851195 4778 generic.go:334] "Generic (PLEG): container finished" podID="67604e51-359f-4c7f-b7df-a4f215a87085" containerID="67c3ac2c335344f6b2ef2e71132a310b8eda046527619858c13389f0ce08da63" exitCode=0 Mar 12 14:04:02 crc kubenswrapper[4778]: I0312 14:04:02.851382 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555404-jwm56" event={"ID":"67604e51-359f-4c7f-b7df-a4f215a87085","Type":"ContainerDied","Data":"67c3ac2c335344f6b2ef2e71132a310b8eda046527619858c13389f0ce08da63"} Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.417945 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.531230 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4zdm\" (UniqueName: \"kubernetes.io/projected/67604e51-359f-4c7f-b7df-a4f215a87085-kube-api-access-j4zdm\") pod \"67604e51-359f-4c7f-b7df-a4f215a87085\" (UID: \"67604e51-359f-4c7f-b7df-a4f215a87085\") " Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.536460 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67604e51-359f-4c7f-b7df-a4f215a87085-kube-api-access-j4zdm" (OuterVolumeSpecName: "kube-api-access-j4zdm") pod "67604e51-359f-4c7f-b7df-a4f215a87085" (UID: "67604e51-359f-4c7f-b7df-a4f215a87085"). InnerVolumeSpecName "kube-api-access-j4zdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.633973 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4zdm\" (UniqueName: \"kubernetes.io/projected/67604e51-359f-4c7f-b7df-a4f215a87085-kube-api-access-j4zdm\") on node \"crc\" DevicePath \"\"" Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.898274 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555404-jwm56" event={"ID":"67604e51-359f-4c7f-b7df-a4f215a87085","Type":"ContainerDied","Data":"1b601cd7deaebbb8c485b6a4bc8cce371566ad54d4f0487a2fe0ff8eb085f00b"} Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.898322 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b601cd7deaebbb8c485b6a4bc8cce371566ad54d4f0487a2fe0ff8eb085f00b" Mar 12 14:04:04 crc kubenswrapper[4778]: I0312 14:04:04.898350 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555404-jwm56" Mar 12 14:04:05 crc kubenswrapper[4778]: I0312 14:04:05.484591 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555398-hhchd"] Mar 12 14:04:05 crc kubenswrapper[4778]: I0312 14:04:05.494064 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555398-hhchd"] Mar 12 14:04:06 crc kubenswrapper[4778]: I0312 14:04:06.265735 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a479324-f9a1-4095-a0b1-7c22fc72eb61" path="/var/lib/kubelet/pods/6a479324-f9a1-4095-a0b1-7c22fc72eb61/volumes" Mar 12 14:04:10 crc kubenswrapper[4778]: I0312 14:04:10.254231 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:04:10 crc kubenswrapper[4778]: E0312 14:04:10.254894 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:04:20 crc kubenswrapper[4778]: I0312 14:04:20.469655 4778 scope.go:117] "RemoveContainer" containerID="7f07e770195234611f35ac5fc4d8c046a4e07dd2e554f881a1c216e51689e210" Mar 12 14:04:22 crc kubenswrapper[4778]: I0312 14:04:22.260421 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:04:22 crc kubenswrapper[4778]: E0312 14:04:22.261899 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:04:34 crc kubenswrapper[4778]: I0312 14:04:34.254053 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:04:35 crc kubenswrapper[4778]: I0312 14:04:35.183424 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"84eb4f64f5e57ea7581e624359f9a06ffee621fbf6407e2f32f007351966b81b"} Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.512919 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-452rn"] Mar 12 14:05:01 crc kubenswrapper[4778]: E0312 14:05:01.513885 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67604e51-359f-4c7f-b7df-a4f215a87085" containerName="oc" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.513899 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="67604e51-359f-4c7f-b7df-a4f215a87085" containerName="oc" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.514120 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="67604e51-359f-4c7f-b7df-a4f215a87085" containerName="oc" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.515655 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.525169 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-452rn"] Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.557837 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-catalog-content\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.558168 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pl6b\" (UniqueName: \"kubernetes.io/projected/3cb5f22f-5751-44aa-9532-51d2f950ee49-kube-api-access-7pl6b\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.558341 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-utilities\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.659986 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-catalog-content\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.660121 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pl6b\" (UniqueName: \"kubernetes.io/projected/3cb5f22f-5751-44aa-9532-51d2f950ee49-kube-api-access-7pl6b\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.660155 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-utilities\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.660607 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-catalog-content\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.660645 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-utilities\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.683018 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pl6b\" (UniqueName: \"kubernetes.io/projected/3cb5f22f-5751-44aa-9532-51d2f950ee49-kube-api-access-7pl6b\") pod \"certified-operators-452rn\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:01 crc kubenswrapper[4778]: I0312 14:05:01.835862 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.370155 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-452rn"] Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.515164 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-79r2g"] Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.517971 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.534431 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-79r2g"] Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.559899 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerStarted","Data":"4c152f196c68c9f25ad019e5ad6afcb88abbcc0e6d071be4e2c4f328dcfe1a46"} Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.584585 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-utilities\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.587284 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6kfs\" (UniqueName: \"kubernetes.io/projected/799d5e51-87fe-402d-9189-d2c430d8225c-kube-api-access-n6kfs\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.587502 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-catalog-content\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.689065 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6kfs\" (UniqueName: \"kubernetes.io/projected/799d5e51-87fe-402d-9189-d2c430d8225c-kube-api-access-n6kfs\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.689170 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-catalog-content\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.689303 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-utilities\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.689676 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-catalog-content\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.689771 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-utilities\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.714321 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6kfs\" (UniqueName: \"kubernetes.io/projected/799d5e51-87fe-402d-9189-d2c430d8225c-kube-api-access-n6kfs\") pod \"community-operators-79r2g\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:02 crc kubenswrapper[4778]: I0312 14:05:02.878417 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:03 crc kubenswrapper[4778]: I0312 14:05:03.404088 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-79r2g"] Mar 12 14:05:03 crc kubenswrapper[4778]: I0312 14:05:03.570623 4778 generic.go:334] "Generic (PLEG): container finished" podID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerID="fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b" exitCode=0 Mar 12 14:05:03 crc kubenswrapper[4778]: I0312 14:05:03.570807 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerDied","Data":"fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b"} Mar 12 14:05:03 crc kubenswrapper[4778]: I0312 14:05:03.572277 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerStarted","Data":"24c962fb193a47f0c7153106b10da358aafd6cfe8775f1e11077c86cd86f8779"} Mar 12 14:05:04 crc kubenswrapper[4778]: I0312 14:05:04.583463 4778 generic.go:334] "Generic (PLEG): container finished" podID="799d5e51-87fe-402d-9189-d2c430d8225c" containerID="5311e6c99490119871346b089342ebf7ab5a0ba98c0dc3500c748eb884a5609f" exitCode=0 Mar 12 14:05:04 crc kubenswrapper[4778]: I0312 14:05:04.583552 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerDied","Data":"5311e6c99490119871346b089342ebf7ab5a0ba98c0dc3500c748eb884a5609f"} Mar 12 14:05:05 crc kubenswrapper[4778]: I0312 14:05:05.598941 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerStarted","Data":"6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60"} Mar 12 14:05:06 crc kubenswrapper[4778]: I0312 14:05:06.609760 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerStarted","Data":"df1ceb7dfdb20eb8574a75ba7a0b1ae7824eb1149a2e784e4f635dced47b8167"} Mar 12 14:05:08 crc kubenswrapper[4778]: I0312 14:05:08.680789 4778 generic.go:334] "Generic (PLEG): container finished" podID="799d5e51-87fe-402d-9189-d2c430d8225c" containerID="df1ceb7dfdb20eb8574a75ba7a0b1ae7824eb1149a2e784e4f635dced47b8167" exitCode=0 Mar 12 14:05:08 crc kubenswrapper[4778]: I0312 14:05:08.680887 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerDied","Data":"df1ceb7dfdb20eb8574a75ba7a0b1ae7824eb1149a2e784e4f635dced47b8167"} Mar 12 14:05:08 crc kubenswrapper[4778]: I0312 14:05:08.683889 4778 generic.go:334] "Generic (PLEG): container finished" podID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerID="6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60" exitCode=0 Mar 12 14:05:08 crc kubenswrapper[4778]: I0312 14:05:08.683920 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerDied","Data":"6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60"} Mar 12 14:05:09 crc kubenswrapper[4778]: I0312 14:05:09.697007 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerStarted","Data":"27772db6c5a7076fc1876873c5a4dffb045dd74854ff2e79c70cca97cebb207f"} Mar 12 14:05:09 crc kubenswrapper[4778]: I0312 14:05:09.700160 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerStarted","Data":"de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d"} Mar 12 14:05:09 crc kubenswrapper[4778]: I0312 14:05:09.719556 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-79r2g" podStartSLOduration=3.165486016 podStartE2EDuration="7.719535819s" podCreationTimestamp="2026-03-12 14:05:02 +0000 UTC" firstStartedPulling="2026-03-12 14:05:04.585838231 +0000 UTC m=+3323.034533627" lastFinishedPulling="2026-03-12 14:05:09.139888034 +0000 UTC m=+3327.588583430" observedRunningTime="2026-03-12 14:05:09.713839367 +0000 UTC m=+3328.162534763" watchObservedRunningTime="2026-03-12 14:05:09.719535819 +0000 UTC m=+3328.168231215" Mar 12 14:05:11 crc kubenswrapper[4778]: I0312 14:05:11.836908 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:11 crc kubenswrapper[4778]: I0312 14:05:11.837595 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:12 crc kubenswrapper[4778]: I0312 14:05:12.879252 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:12 crc kubenswrapper[4778]: I0312 14:05:12.879606 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:12 crc kubenswrapper[4778]: I0312 14:05:12.880981 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-452rn" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="registry-server" probeResult="failure" output=< Mar 12 14:05:12 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:05:12 crc kubenswrapper[4778]: > Mar 12 14:05:12 crc kubenswrapper[4778]: I0312 14:05:12.925387 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:12 crc kubenswrapper[4778]: I0312 14:05:12.950260 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-452rn" podStartSLOduration=6.272736369 podStartE2EDuration="11.950231393s" podCreationTimestamp="2026-03-12 14:05:01 +0000 UTC" firstStartedPulling="2026-03-12 14:05:03.573743273 +0000 UTC m=+3322.022438669" lastFinishedPulling="2026-03-12 14:05:09.251238297 +0000 UTC m=+3327.699933693" observedRunningTime="2026-03-12 14:05:09.734456443 +0000 UTC m=+3328.183151859" watchObservedRunningTime="2026-03-12 14:05:12.950231393 +0000 UTC m=+3331.398926789" Mar 12 14:05:21 crc kubenswrapper[4778]: I0312 14:05:21.914634 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:21 crc kubenswrapper[4778]: I0312 14:05:21.998812 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:22 crc kubenswrapper[4778]: I0312 14:05:22.172087 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-452rn"] Mar 12 14:05:22 crc kubenswrapper[4778]: I0312 14:05:22.924492 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:23 crc kubenswrapper[4778]: I0312 14:05:23.917352 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-452rn" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="registry-server" containerID="cri-o://de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d" gracePeriod=2 Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.569934 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.570375 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-79r2g"] Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.570596 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-79r2g" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="registry-server" containerID="cri-o://27772db6c5a7076fc1876873c5a4dffb045dd74854ff2e79c70cca97cebb207f" gracePeriod=2 Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.669247 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-catalog-content\") pod \"3cb5f22f-5751-44aa-9532-51d2f950ee49\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.669455 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-utilities\") pod \"3cb5f22f-5751-44aa-9532-51d2f950ee49\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.669510 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pl6b\" (UniqueName: \"kubernetes.io/projected/3cb5f22f-5751-44aa-9532-51d2f950ee49-kube-api-access-7pl6b\") pod \"3cb5f22f-5751-44aa-9532-51d2f950ee49\" (UID: \"3cb5f22f-5751-44aa-9532-51d2f950ee49\") " Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.670505 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-utilities" (OuterVolumeSpecName: "utilities") pod "3cb5f22f-5751-44aa-9532-51d2f950ee49" (UID: "3cb5f22f-5751-44aa-9532-51d2f950ee49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.699281 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb5f22f-5751-44aa-9532-51d2f950ee49-kube-api-access-7pl6b" (OuterVolumeSpecName: "kube-api-access-7pl6b") pod "3cb5f22f-5751-44aa-9532-51d2f950ee49" (UID: "3cb5f22f-5751-44aa-9532-51d2f950ee49"). InnerVolumeSpecName "kube-api-access-7pl6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.743516 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cb5f22f-5751-44aa-9532-51d2f950ee49" (UID: "3cb5f22f-5751-44aa-9532-51d2f950ee49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.771303 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pl6b\" (UniqueName: \"kubernetes.io/projected/3cb5f22f-5751-44aa-9532-51d2f950ee49-kube-api-access-7pl6b\") on node \"crc\" DevicePath \"\"" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.771333 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.771346 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cb5f22f-5751-44aa-9532-51d2f950ee49-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.931263 4778 generic.go:334] "Generic (PLEG): container finished" podID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerID="de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d" exitCode=0 Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.931364 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-452rn" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.931384 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerDied","Data":"de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d"} Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.931458 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-452rn" event={"ID":"3cb5f22f-5751-44aa-9532-51d2f950ee49","Type":"ContainerDied","Data":"4c152f196c68c9f25ad019e5ad6afcb88abbcc0e6d071be4e2c4f328dcfe1a46"} Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.931486 4778 scope.go:117] "RemoveContainer" containerID="de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d" Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.934668 4778 generic.go:334] "Generic (PLEG): container finished" podID="799d5e51-87fe-402d-9189-d2c430d8225c" containerID="27772db6c5a7076fc1876873c5a4dffb045dd74854ff2e79c70cca97cebb207f" exitCode=0 Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.934703 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerDied","Data":"27772db6c5a7076fc1876873c5a4dffb045dd74854ff2e79c70cca97cebb207f"} Mar 12 14:05:24 crc kubenswrapper[4778]: I0312 14:05:24.963747 4778 scope.go:117] "RemoveContainer" containerID="6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.011100 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-452rn"] Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.020555 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-452rn"] Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.047639 4778 scope.go:117] "RemoveContainer" containerID="fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.095867 4778 scope.go:117] "RemoveContainer" containerID="de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d" Mar 12 14:05:25 crc kubenswrapper[4778]: E0312 14:05:25.096746 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d\": container with ID starting with de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d not found: ID does not exist" containerID="de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.096815 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d"} err="failed to get container status \"de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d\": rpc error: code = NotFound desc = could not find container \"de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d\": container with ID starting with de577889462420ffee500a8838f5146588348b4a1a1dbbd50304bd68fabfdd7d not found: ID does not exist" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.096845 4778 scope.go:117] "RemoveContainer" containerID="6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60" Mar 12 14:05:25 crc kubenswrapper[4778]: E0312 14:05:25.097329 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60\": container with ID starting with 6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60 not found: ID does not exist" containerID="6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.097352 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60"} err="failed to get container status \"6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60\": rpc error: code = NotFound desc = could not find container \"6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60\": container with ID starting with 6089ff4895f75a6eaae315a037e64eff2551aa8e0e865f56674c8421758bee60 not found: ID does not exist" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.097370 4778 scope.go:117] "RemoveContainer" containerID="fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b" Mar 12 14:05:25 crc kubenswrapper[4778]: E0312 14:05:25.098308 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b\": container with ID starting with fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b not found: ID does not exist" containerID="fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.098334 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b"} err="failed to get container status \"fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b\": rpc error: code = NotFound desc = could not find container \"fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b\": container with ID starting with fe41bd24a0ab71cebcec1476a92f57f89206d945a71642088676557cf061ae1b not found: ID does not exist" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.306434 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.395638 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-utilities\") pod \"799d5e51-87fe-402d-9189-d2c430d8225c\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.395855 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6kfs\" (UniqueName: \"kubernetes.io/projected/799d5e51-87fe-402d-9189-d2c430d8225c-kube-api-access-n6kfs\") pod \"799d5e51-87fe-402d-9189-d2c430d8225c\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.395995 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-catalog-content\") pod \"799d5e51-87fe-402d-9189-d2c430d8225c\" (UID: \"799d5e51-87fe-402d-9189-d2c430d8225c\") " Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.396650 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-utilities" (OuterVolumeSpecName: "utilities") pod "799d5e51-87fe-402d-9189-d2c430d8225c" (UID: "799d5e51-87fe-402d-9189-d2c430d8225c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.396933 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.409639 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/799d5e51-87fe-402d-9189-d2c430d8225c-kube-api-access-n6kfs" (OuterVolumeSpecName: "kube-api-access-n6kfs") pod "799d5e51-87fe-402d-9189-d2c430d8225c" (UID: "799d5e51-87fe-402d-9189-d2c430d8225c"). InnerVolumeSpecName "kube-api-access-n6kfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.453880 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "799d5e51-87fe-402d-9189-d2c430d8225c" (UID: "799d5e51-87fe-402d-9189-d2c430d8225c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.499459 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6kfs\" (UniqueName: \"kubernetes.io/projected/799d5e51-87fe-402d-9189-d2c430d8225c-kube-api-access-n6kfs\") on node \"crc\" DevicePath \"\"" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.499508 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799d5e51-87fe-402d-9189-d2c430d8225c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.950265 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79r2g" event={"ID":"799d5e51-87fe-402d-9189-d2c430d8225c","Type":"ContainerDied","Data":"24c962fb193a47f0c7153106b10da358aafd6cfe8775f1e11077c86cd86f8779"} Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.950328 4778 scope.go:117] "RemoveContainer" containerID="27772db6c5a7076fc1876873c5a4dffb045dd74854ff2e79c70cca97cebb207f" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.951341 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79r2g" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.975295 4778 scope.go:117] "RemoveContainer" containerID="df1ceb7dfdb20eb8574a75ba7a0b1ae7824eb1149a2e784e4f635dced47b8167" Mar 12 14:05:25 crc kubenswrapper[4778]: I0312 14:05:25.998813 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-79r2g"] Mar 12 14:05:26 crc kubenswrapper[4778]: I0312 14:05:26.007736 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-79r2g"] Mar 12 14:05:26 crc kubenswrapper[4778]: I0312 14:05:26.007791 4778 scope.go:117] "RemoveContainer" containerID="5311e6c99490119871346b089342ebf7ab5a0ba98c0dc3500c748eb884a5609f" Mar 12 14:05:26 crc kubenswrapper[4778]: I0312 14:05:26.266154 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" path="/var/lib/kubelet/pods/3cb5f22f-5751-44aa-9532-51d2f950ee49/volumes" Mar 12 14:05:26 crc kubenswrapper[4778]: I0312 14:05:26.267299 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" path="/var/lib/kubelet/pods/799d5e51-87fe-402d-9189-d2c430d8225c/volumes" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.159065 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555406-44v2c"] Mar 12 14:06:00 crc kubenswrapper[4778]: E0312 14:06:00.160089 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="extract-utilities" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160104 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="extract-utilities" Mar 12 14:06:00 crc kubenswrapper[4778]: E0312 14:06:00.160118 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="registry-server" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160125 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="registry-server" Mar 12 14:06:00 crc kubenswrapper[4778]: E0312 14:06:00.160134 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="registry-server" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160142 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="registry-server" Mar 12 14:06:00 crc kubenswrapper[4778]: E0312 14:06:00.160163 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="extract-content" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160169 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="extract-content" Mar 12 14:06:00 crc kubenswrapper[4778]: E0312 14:06:00.160177 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="extract-utilities" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160199 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="extract-utilities" Mar 12 14:06:00 crc kubenswrapper[4778]: E0312 14:06:00.160212 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="extract-content" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160217 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="extract-content" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160415 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb5f22f-5751-44aa-9532-51d2f950ee49" containerName="registry-server" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.160438 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="799d5e51-87fe-402d-9189-d2c430d8225c" containerName="registry-server" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.161122 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.165215 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.165530 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.167554 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.169730 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555406-44v2c"] Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.255227 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dsrm\" (UniqueName: \"kubernetes.io/projected/2e7c143c-4173-450a-afa1-587a3927f2d4-kube-api-access-2dsrm\") pod \"auto-csr-approver-29555406-44v2c\" (UID: \"2e7c143c-4173-450a-afa1-587a3927f2d4\") " pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.357715 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dsrm\" (UniqueName: \"kubernetes.io/projected/2e7c143c-4173-450a-afa1-587a3927f2d4-kube-api-access-2dsrm\") pod \"auto-csr-approver-29555406-44v2c\" (UID: \"2e7c143c-4173-450a-afa1-587a3927f2d4\") " pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.387593 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dsrm\" (UniqueName: \"kubernetes.io/projected/2e7c143c-4173-450a-afa1-587a3927f2d4-kube-api-access-2dsrm\") pod \"auto-csr-approver-29555406-44v2c\" (UID: \"2e7c143c-4173-450a-afa1-587a3927f2d4\") " pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.481653 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:00 crc kubenswrapper[4778]: I0312 14:06:00.967645 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555406-44v2c"] Mar 12 14:06:01 crc kubenswrapper[4778]: I0312 14:06:01.262198 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555406-44v2c" event={"ID":"2e7c143c-4173-450a-afa1-587a3927f2d4","Type":"ContainerStarted","Data":"0369ea96b618e30848880d0734baef647a7b442bbb78ff57eb0d051a016f4603"} Mar 12 14:06:03 crc kubenswrapper[4778]: I0312 14:06:03.279811 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555406-44v2c" event={"ID":"2e7c143c-4173-450a-afa1-587a3927f2d4","Type":"ContainerStarted","Data":"2c9bf5717fd9b2c8602b788cdc193d4c283ef18c6a74310ea29b1e044df19e27"} Mar 12 14:06:03 crc kubenswrapper[4778]: I0312 14:06:03.296094 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555406-44v2c" podStartSLOduration=1.401232633 podStartE2EDuration="3.296079335s" podCreationTimestamp="2026-03-12 14:06:00 +0000 UTC" firstStartedPulling="2026-03-12 14:06:00.959871276 +0000 UTC m=+3379.408566672" lastFinishedPulling="2026-03-12 14:06:02.854717978 +0000 UTC m=+3381.303413374" observedRunningTime="2026-03-12 14:06:03.292547554 +0000 UTC m=+3381.741242950" watchObservedRunningTime="2026-03-12 14:06:03.296079335 +0000 UTC m=+3381.744774731" Mar 12 14:06:04 crc kubenswrapper[4778]: I0312 14:06:04.306934 4778 generic.go:334] "Generic (PLEG): container finished" podID="2e7c143c-4173-450a-afa1-587a3927f2d4" containerID="2c9bf5717fd9b2c8602b788cdc193d4c283ef18c6a74310ea29b1e044df19e27" exitCode=0 Mar 12 14:06:04 crc kubenswrapper[4778]: I0312 14:06:04.307232 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555406-44v2c" event={"ID":"2e7c143c-4173-450a-afa1-587a3927f2d4","Type":"ContainerDied","Data":"2c9bf5717fd9b2c8602b788cdc193d4c283ef18c6a74310ea29b1e044df19e27"} Mar 12 14:06:05 crc kubenswrapper[4778]: I0312 14:06:05.850681 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:05 crc kubenswrapper[4778]: I0312 14:06:05.982158 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dsrm\" (UniqueName: \"kubernetes.io/projected/2e7c143c-4173-450a-afa1-587a3927f2d4-kube-api-access-2dsrm\") pod \"2e7c143c-4173-450a-afa1-587a3927f2d4\" (UID: \"2e7c143c-4173-450a-afa1-587a3927f2d4\") " Mar 12 14:06:05 crc kubenswrapper[4778]: I0312 14:06:05.999395 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e7c143c-4173-450a-afa1-587a3927f2d4-kube-api-access-2dsrm" (OuterVolumeSpecName: "kube-api-access-2dsrm") pod "2e7c143c-4173-450a-afa1-587a3927f2d4" (UID: "2e7c143c-4173-450a-afa1-587a3927f2d4"). InnerVolumeSpecName "kube-api-access-2dsrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:06:06 crc kubenswrapper[4778]: I0312 14:06:06.086492 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dsrm\" (UniqueName: \"kubernetes.io/projected/2e7c143c-4173-450a-afa1-587a3927f2d4-kube-api-access-2dsrm\") on node \"crc\" DevicePath \"\"" Mar 12 14:06:06 crc kubenswrapper[4778]: I0312 14:06:06.325389 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555406-44v2c" event={"ID":"2e7c143c-4173-450a-afa1-587a3927f2d4","Type":"ContainerDied","Data":"0369ea96b618e30848880d0734baef647a7b442bbb78ff57eb0d051a016f4603"} Mar 12 14:06:06 crc kubenswrapper[4778]: I0312 14:06:06.325433 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0369ea96b618e30848880d0734baef647a7b442bbb78ff57eb0d051a016f4603" Mar 12 14:06:06 crc kubenswrapper[4778]: I0312 14:06:06.325460 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555406-44v2c" Mar 12 14:06:06 crc kubenswrapper[4778]: I0312 14:06:06.369847 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555400-c5pzt"] Mar 12 14:06:06 crc kubenswrapper[4778]: I0312 14:06:06.378535 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555400-c5pzt"] Mar 12 14:06:08 crc kubenswrapper[4778]: I0312 14:06:08.265496 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2425d74f-ef53-43bc-8c8f-976333a9cc6a" path="/var/lib/kubelet/pods/2425d74f-ef53-43bc-8c8f-976333a9cc6a/volumes" Mar 12 14:06:20 crc kubenswrapper[4778]: I0312 14:06:20.570611 4778 scope.go:117] "RemoveContainer" containerID="363f3ad00ca01b087e83fcbce9630716537dd1aa2dde624be9a2f51cfec1e8a6" Mar 12 14:06:58 crc kubenswrapper[4778]: I0312 14:06:58.557459 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:06:58 crc kubenswrapper[4778]: I0312 14:06:58.558022 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.068231 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9fgw2"] Mar 12 14:07:04 crc kubenswrapper[4778]: E0312 14:07:04.069464 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7c143c-4173-450a-afa1-587a3927f2d4" containerName="oc" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.069500 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7c143c-4173-450a-afa1-587a3927f2d4" containerName="oc" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.069788 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e7c143c-4173-450a-afa1-587a3927f2d4" containerName="oc" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.073669 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.108555 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fgw2"] Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.196636 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-utilities\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.196718 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stkwf\" (UniqueName: \"kubernetes.io/projected/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-kube-api-access-stkwf\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.196788 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-catalog-content\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.299098 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-utilities\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.299324 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stkwf\" (UniqueName: \"kubernetes.io/projected/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-kube-api-access-stkwf\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.299520 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-catalog-content\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.300036 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-utilities\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.300819 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-catalog-content\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.322442 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stkwf\" (UniqueName: \"kubernetes.io/projected/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-kube-api-access-stkwf\") pod \"redhat-operators-9fgw2\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.404519 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.892433 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fgw2"] Mar 12 14:07:04 crc kubenswrapper[4778]: I0312 14:07:04.946025 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerStarted","Data":"c323eb083241a5c46e1a35088e34c27b0835efe1120e3b7aa0048e10bce8bed7"} Mar 12 14:07:05 crc kubenswrapper[4778]: I0312 14:07:05.960624 4778 generic.go:334] "Generic (PLEG): container finished" podID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerID="c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d" exitCode=0 Mar 12 14:07:05 crc kubenswrapper[4778]: I0312 14:07:05.960966 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerDied","Data":"c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d"} Mar 12 14:07:05 crc kubenswrapper[4778]: I0312 14:07:05.965041 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:07:07 crc kubenswrapper[4778]: I0312 14:07:07.981997 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerStarted","Data":"cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683"} Mar 12 14:07:10 crc kubenswrapper[4778]: I0312 14:07:09.999720 4778 generic.go:334] "Generic (PLEG): container finished" podID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerID="cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683" exitCode=0 Mar 12 14:07:10 crc kubenswrapper[4778]: I0312 14:07:09.999886 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerDied","Data":"cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683"} Mar 12 14:07:11 crc kubenswrapper[4778]: I0312 14:07:11.012736 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerStarted","Data":"13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0"} Mar 12 14:07:11 crc kubenswrapper[4778]: I0312 14:07:11.036474 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9fgw2" podStartSLOduration=2.358519127 podStartE2EDuration="7.036444774s" podCreationTimestamp="2026-03-12 14:07:04 +0000 UTC" firstStartedPulling="2026-03-12 14:07:05.964068828 +0000 UTC m=+3444.412764224" lastFinishedPulling="2026-03-12 14:07:10.641994475 +0000 UTC m=+3449.090689871" observedRunningTime="2026-03-12 14:07:11.035098796 +0000 UTC m=+3449.483794202" watchObservedRunningTime="2026-03-12 14:07:11.036444774 +0000 UTC m=+3449.485140170" Mar 12 14:07:14 crc kubenswrapper[4778]: I0312 14:07:14.404723 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:14 crc kubenswrapper[4778]: I0312 14:07:14.405892 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:15 crc kubenswrapper[4778]: I0312 14:07:15.459565 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9fgw2" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="registry-server" probeResult="failure" output=< Mar 12 14:07:15 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:07:15 crc kubenswrapper[4778]: > Mar 12 14:07:24 crc kubenswrapper[4778]: I0312 14:07:24.456917 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:24 crc kubenswrapper[4778]: I0312 14:07:24.508827 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:24 crc kubenswrapper[4778]: I0312 14:07:24.702019 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fgw2"] Mar 12 14:07:26 crc kubenswrapper[4778]: I0312 14:07:26.145907 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9fgw2" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="registry-server" containerID="cri-o://13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0" gracePeriod=2 Mar 12 14:07:26 crc kubenswrapper[4778]: I0312 14:07:26.806625 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:26 crc kubenswrapper[4778]: I0312 14:07:26.997622 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stkwf\" (UniqueName: \"kubernetes.io/projected/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-kube-api-access-stkwf\") pod \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " Mar 12 14:07:26 crc kubenswrapper[4778]: I0312 14:07:26.997780 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-utilities\") pod \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " Mar 12 14:07:26 crc kubenswrapper[4778]: I0312 14:07:26.997800 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-catalog-content\") pod \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\" (UID: \"1169365c-cb69-43bd-9a4d-fcc7c2a467e1\") " Mar 12 14:07:26 crc kubenswrapper[4778]: I0312 14:07:26.999718 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-utilities" (OuterVolumeSpecName: "utilities") pod "1169365c-cb69-43bd-9a4d-fcc7c2a467e1" (UID: "1169365c-cb69-43bd-9a4d-fcc7c2a467e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.007604 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-kube-api-access-stkwf" (OuterVolumeSpecName: "kube-api-access-stkwf") pod "1169365c-cb69-43bd-9a4d-fcc7c2a467e1" (UID: "1169365c-cb69-43bd-9a4d-fcc7c2a467e1"). InnerVolumeSpecName "kube-api-access-stkwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.100324 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stkwf\" (UniqueName: \"kubernetes.io/projected/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-kube-api-access-stkwf\") on node \"crc\" DevicePath \"\"" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.100357 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.138013 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1169365c-cb69-43bd-9a4d-fcc7c2a467e1" (UID: "1169365c-cb69-43bd-9a4d-fcc7c2a467e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.157015 4778 generic.go:334] "Generic (PLEG): container finished" podID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerID="13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0" exitCode=0 Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.157058 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerDied","Data":"13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0"} Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.157085 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fgw2" event={"ID":"1169365c-cb69-43bd-9a4d-fcc7c2a467e1","Type":"ContainerDied","Data":"c323eb083241a5c46e1a35088e34c27b0835efe1120e3b7aa0048e10bce8bed7"} Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.157106 4778 scope.go:117] "RemoveContainer" containerID="13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.157114 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fgw2" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.186412 4778 scope.go:117] "RemoveContainer" containerID="cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.201548 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1169365c-cb69-43bd-9a4d-fcc7c2a467e1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.211957 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fgw2"] Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.217477 4778 scope.go:117] "RemoveContainer" containerID="c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.222083 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9fgw2"] Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.254951 4778 scope.go:117] "RemoveContainer" containerID="13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0" Mar 12 14:07:27 crc kubenswrapper[4778]: E0312 14:07:27.255411 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0\": container with ID starting with 13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0 not found: ID does not exist" containerID="13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.255455 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0"} err="failed to get container status \"13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0\": rpc error: code = NotFound desc = could not find container \"13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0\": container with ID starting with 13fc93c1394603c467ef3707a22e1e5c3b4a3d019842a6cf1758b714f1863da0 not found: ID does not exist" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.255477 4778 scope.go:117] "RemoveContainer" containerID="cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683" Mar 12 14:07:27 crc kubenswrapper[4778]: E0312 14:07:27.256531 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683\": container with ID starting with cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683 not found: ID does not exist" containerID="cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.256573 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683"} err="failed to get container status \"cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683\": rpc error: code = NotFound desc = could not find container \"cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683\": container with ID starting with cb2c67312712b020bfef441f944b478e1c4b6ca687e4ed1e7e31fd2401b71683 not found: ID does not exist" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.256604 4778 scope.go:117] "RemoveContainer" containerID="c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d" Mar 12 14:07:27 crc kubenswrapper[4778]: E0312 14:07:27.256907 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d\": container with ID starting with c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d not found: ID does not exist" containerID="c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d" Mar 12 14:07:27 crc kubenswrapper[4778]: I0312 14:07:27.256934 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d"} err="failed to get container status \"c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d\": rpc error: code = NotFound desc = could not find container \"c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d\": container with ID starting with c1a67074c23faef29e037065d8226fda81b2b6cc308527fda75db7c9667ac65d not found: ID does not exist" Mar 12 14:07:28 crc kubenswrapper[4778]: I0312 14:07:28.264974 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" path="/var/lib/kubelet/pods/1169365c-cb69-43bd-9a4d-fcc7c2a467e1/volumes" Mar 12 14:07:28 crc kubenswrapper[4778]: I0312 14:07:28.558167 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:07:28 crc kubenswrapper[4778]: I0312 14:07:28.558341 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:07:58 crc kubenswrapper[4778]: I0312 14:07:58.557613 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:07:58 crc kubenswrapper[4778]: I0312 14:07:58.558176 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:07:58 crc kubenswrapper[4778]: I0312 14:07:58.558265 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:07:58 crc kubenswrapper[4778]: I0312 14:07:58.559118 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"84eb4f64f5e57ea7581e624359f9a06ffee621fbf6407e2f32f007351966b81b"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:07:58 crc kubenswrapper[4778]: I0312 14:07:58.559181 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://84eb4f64f5e57ea7581e624359f9a06ffee621fbf6407e2f32f007351966b81b" gracePeriod=600 Mar 12 14:07:59 crc kubenswrapper[4778]: I0312 14:07:59.444841 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="84eb4f64f5e57ea7581e624359f9a06ffee621fbf6407e2f32f007351966b81b" exitCode=0 Mar 12 14:07:59 crc kubenswrapper[4778]: I0312 14:07:59.445124 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"84eb4f64f5e57ea7581e624359f9a06ffee621fbf6407e2f32f007351966b81b"} Mar 12 14:07:59 crc kubenswrapper[4778]: I0312 14:07:59.445156 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2"} Mar 12 14:07:59 crc kubenswrapper[4778]: I0312 14:07:59.445176 4778 scope.go:117] "RemoveContainer" containerID="264800b09f45ccd4290c89a1d8ecad1ba09b58524e636d065df86104736d56c0" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.152748 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555408-92kzm"] Mar 12 14:08:00 crc kubenswrapper[4778]: E0312 14:08:00.153837 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="registry-server" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.153859 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="registry-server" Mar 12 14:08:00 crc kubenswrapper[4778]: E0312 14:08:00.153876 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="extract-content" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.153887 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="extract-content" Mar 12 14:08:00 crc kubenswrapper[4778]: E0312 14:08:00.153920 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="extract-utilities" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.153933 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="extract-utilities" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.154610 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1169365c-cb69-43bd-9a4d-fcc7c2a467e1" containerName="registry-server" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.155928 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.158597 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.158598 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.164258 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555408-92kzm"] Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.175109 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.339029 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb6lk\" (UniqueName: \"kubernetes.io/projected/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2-kube-api-access-vb6lk\") pod \"auto-csr-approver-29555408-92kzm\" (UID: \"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2\") " pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.441537 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb6lk\" (UniqueName: \"kubernetes.io/projected/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2-kube-api-access-vb6lk\") pod \"auto-csr-approver-29555408-92kzm\" (UID: \"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2\") " pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.463209 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb6lk\" (UniqueName: \"kubernetes.io/projected/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2-kube-api-access-vb6lk\") pod \"auto-csr-approver-29555408-92kzm\" (UID: \"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2\") " pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.488081 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:00 crc kubenswrapper[4778]: I0312 14:08:00.987835 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555408-92kzm"] Mar 12 14:08:01 crc kubenswrapper[4778]: I0312 14:08:01.465638 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555408-92kzm" event={"ID":"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2","Type":"ContainerStarted","Data":"9319c323739957e4e4b497353bc14e4abe0ae389088d65b7d2ee4703ab69b7fe"} Mar 12 14:08:02 crc kubenswrapper[4778]: I0312 14:08:02.474593 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555408-92kzm" event={"ID":"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2","Type":"ContainerStarted","Data":"9557c198d563e8bc3c1bd0c3db7f0caaf03fda89ffb3294270c9a1e6bbdd5486"} Mar 12 14:08:03 crc kubenswrapper[4778]: I0312 14:08:03.484199 4778 generic.go:334] "Generic (PLEG): container finished" podID="f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2" containerID="9557c198d563e8bc3c1bd0c3db7f0caaf03fda89ffb3294270c9a1e6bbdd5486" exitCode=0 Mar 12 14:08:03 crc kubenswrapper[4778]: I0312 14:08:03.484302 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555408-92kzm" event={"ID":"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2","Type":"ContainerDied","Data":"9557c198d563e8bc3c1bd0c3db7f0caaf03fda89ffb3294270c9a1e6bbdd5486"} Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.163023 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.237012 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb6lk\" (UniqueName: \"kubernetes.io/projected/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2-kube-api-access-vb6lk\") pod \"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2\" (UID: \"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2\") " Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.243886 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2-kube-api-access-vb6lk" (OuterVolumeSpecName: "kube-api-access-vb6lk") pod "f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2" (UID: "f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2"). InnerVolumeSpecName "kube-api-access-vb6lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.339020 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb6lk\" (UniqueName: \"kubernetes.io/projected/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2-kube-api-access-vb6lk\") on node \"crc\" DevicePath \"\"" Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.354841 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555402-xtt9v"] Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.367574 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555402-xtt9v"] Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.511199 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555408-92kzm" event={"ID":"f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2","Type":"ContainerDied","Data":"9319c323739957e4e4b497353bc14e4abe0ae389088d65b7d2ee4703ab69b7fe"} Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.511243 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9319c323739957e4e4b497353bc14e4abe0ae389088d65b7d2ee4703ab69b7fe" Mar 12 14:08:05 crc kubenswrapper[4778]: I0312 14:08:05.511252 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555408-92kzm" Mar 12 14:08:06 crc kubenswrapper[4778]: I0312 14:08:06.268218 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be917952-7177-4ef5-9efa-7858d1a11ded" path="/var/lib/kubelet/pods/be917952-7177-4ef5-9efa-7858d1a11ded/volumes" Mar 12 14:08:20 crc kubenswrapper[4778]: I0312 14:08:20.700114 4778 scope.go:117] "RemoveContainer" containerID="b626545edbe9764de0b916e68f0836b92c6dbff05d2ae4f9ae924f063217aca7" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.549775 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hlz27"] Mar 12 14:08:58 crc kubenswrapper[4778]: E0312 14:08:58.550897 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2" containerName="oc" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.550915 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2" containerName="oc" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.551158 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2" containerName="oc" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.553103 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.587809 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlz27"] Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.702250 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-catalog-content\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.702414 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skwjz\" (UniqueName: \"kubernetes.io/projected/6874a57b-b8de-4009-972e-c2c6d0635745-kube-api-access-skwjz\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.702445 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-utilities\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.804842 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skwjz\" (UniqueName: \"kubernetes.io/projected/6874a57b-b8de-4009-972e-c2c6d0635745-kube-api-access-skwjz\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.804900 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-utilities\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.804985 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-catalog-content\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.805519 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-catalog-content\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.805592 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-utilities\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.828453 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skwjz\" (UniqueName: \"kubernetes.io/projected/6874a57b-b8de-4009-972e-c2c6d0635745-kube-api-access-skwjz\") pod \"redhat-marketplace-hlz27\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:58 crc kubenswrapper[4778]: I0312 14:08:58.899635 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:08:59 crc kubenswrapper[4778]: I0312 14:08:59.438272 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlz27"] Mar 12 14:09:00 crc kubenswrapper[4778]: I0312 14:09:00.017284 4778 generic.go:334] "Generic (PLEG): container finished" podID="6874a57b-b8de-4009-972e-c2c6d0635745" containerID="98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f" exitCode=0 Mar 12 14:09:00 crc kubenswrapper[4778]: I0312 14:09:00.017349 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerDied","Data":"98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f"} Mar 12 14:09:00 crc kubenswrapper[4778]: I0312 14:09:00.017567 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerStarted","Data":"3d3144763f7f3b537cab383b60e4a1735726bba4d562907ab1bda5ecb2039f40"} Mar 12 14:09:02 crc kubenswrapper[4778]: I0312 14:09:02.038173 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerStarted","Data":"a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f"} Mar 12 14:09:03 crc kubenswrapper[4778]: I0312 14:09:03.048544 4778 generic.go:334] "Generic (PLEG): container finished" podID="6874a57b-b8de-4009-972e-c2c6d0635745" containerID="a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f" exitCode=0 Mar 12 14:09:03 crc kubenswrapper[4778]: I0312 14:09:03.048593 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerDied","Data":"a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f"} Mar 12 14:09:04 crc kubenswrapper[4778]: I0312 14:09:04.063584 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerStarted","Data":"2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0"} Mar 12 14:09:04 crc kubenswrapper[4778]: I0312 14:09:04.091006 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hlz27" podStartSLOduration=2.620586221 podStartE2EDuration="6.090987177s" podCreationTimestamp="2026-03-12 14:08:58 +0000 UTC" firstStartedPulling="2026-03-12 14:09:00.019442895 +0000 UTC m=+3558.468138291" lastFinishedPulling="2026-03-12 14:09:03.489843851 +0000 UTC m=+3561.938539247" observedRunningTime="2026-03-12 14:09:04.08474848 +0000 UTC m=+3562.533443876" watchObservedRunningTime="2026-03-12 14:09:04.090987177 +0000 UTC m=+3562.539682583" Mar 12 14:09:08 crc kubenswrapper[4778]: I0312 14:09:08.899875 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:09:08 crc kubenswrapper[4778]: I0312 14:09:08.900340 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:09:08 crc kubenswrapper[4778]: I0312 14:09:08.941701 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:09:09 crc kubenswrapper[4778]: I0312 14:09:09.144233 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:09:09 crc kubenswrapper[4778]: I0312 14:09:09.197949 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlz27"] Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.123559 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hlz27" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="registry-server" containerID="cri-o://2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0" gracePeriod=2 Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.868099 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.954588 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skwjz\" (UniqueName: \"kubernetes.io/projected/6874a57b-b8de-4009-972e-c2c6d0635745-kube-api-access-skwjz\") pod \"6874a57b-b8de-4009-972e-c2c6d0635745\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.954868 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-catalog-content\") pod \"6874a57b-b8de-4009-972e-c2c6d0635745\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.955006 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-utilities\") pod \"6874a57b-b8de-4009-972e-c2c6d0635745\" (UID: \"6874a57b-b8de-4009-972e-c2c6d0635745\") " Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.956305 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-utilities" (OuterVolumeSpecName: "utilities") pod "6874a57b-b8de-4009-972e-c2c6d0635745" (UID: "6874a57b-b8de-4009-972e-c2c6d0635745"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.966694 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6874a57b-b8de-4009-972e-c2c6d0635745-kube-api-access-skwjz" (OuterVolumeSpecName: "kube-api-access-skwjz") pod "6874a57b-b8de-4009-972e-c2c6d0635745" (UID: "6874a57b-b8de-4009-972e-c2c6d0635745"). InnerVolumeSpecName "kube-api-access-skwjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:09:11 crc kubenswrapper[4778]: I0312 14:09:11.999251 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6874a57b-b8de-4009-972e-c2c6d0635745" (UID: "6874a57b-b8de-4009-972e-c2c6d0635745"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.057953 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skwjz\" (UniqueName: \"kubernetes.io/projected/6874a57b-b8de-4009-972e-c2c6d0635745-kube-api-access-skwjz\") on node \"crc\" DevicePath \"\"" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.058016 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.058027 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6874a57b-b8de-4009-972e-c2c6d0635745-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.133451 4778 generic.go:334] "Generic (PLEG): container finished" podID="6874a57b-b8de-4009-972e-c2c6d0635745" containerID="2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0" exitCode=0 Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.134272 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerDied","Data":"2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0"} Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.134398 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlz27" event={"ID":"6874a57b-b8de-4009-972e-c2c6d0635745","Type":"ContainerDied","Data":"3d3144763f7f3b537cab383b60e4a1735726bba4d562907ab1bda5ecb2039f40"} Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.134476 4778 scope.go:117] "RemoveContainer" containerID="2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.134682 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlz27" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.169228 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlz27"] Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.173057 4778 scope.go:117] "RemoveContainer" containerID="a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.180830 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlz27"] Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.193107 4778 scope.go:117] "RemoveContainer" containerID="98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.237094 4778 scope.go:117] "RemoveContainer" containerID="2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0" Mar 12 14:09:12 crc kubenswrapper[4778]: E0312 14:09:12.237733 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0\": container with ID starting with 2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0 not found: ID does not exist" containerID="2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.237790 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0"} err="failed to get container status \"2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0\": rpc error: code = NotFound desc = could not find container \"2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0\": container with ID starting with 2db4cd718dbcfe21333c333b517ee99189e4e51f13b211cdd3a2c56aba5558c0 not found: ID does not exist" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.237846 4778 scope.go:117] "RemoveContainer" containerID="a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f" Mar 12 14:09:12 crc kubenswrapper[4778]: E0312 14:09:12.238133 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f\": container with ID starting with a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f not found: ID does not exist" containerID="a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.238156 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f"} err="failed to get container status \"a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f\": rpc error: code = NotFound desc = could not find container \"a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f\": container with ID starting with a32e5ae0574f6d2c567ab56645cb914f613126a04b379308b34a6d869d8e358f not found: ID does not exist" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.238169 4778 scope.go:117] "RemoveContainer" containerID="98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f" Mar 12 14:09:12 crc kubenswrapper[4778]: E0312 14:09:12.238376 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f\": container with ID starting with 98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f not found: ID does not exist" containerID="98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.238396 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f"} err="failed to get container status \"98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f\": rpc error: code = NotFound desc = could not find container \"98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f\": container with ID starting with 98d0a915d740c9246f67fafb55f554a2953e84bea7d0fe5a5f0b921473930f4f not found: ID does not exist" Mar 12 14:09:12 crc kubenswrapper[4778]: I0312 14:09:12.272467 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" path="/var/lib/kubelet/pods/6874a57b-b8de-4009-972e-c2c6d0635745/volumes" Mar 12 14:09:58 crc kubenswrapper[4778]: I0312 14:09:58.558870 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:09:58 crc kubenswrapper[4778]: I0312 14:09:58.560448 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.155311 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555410-ptqps"] Mar 12 14:10:00 crc kubenswrapper[4778]: E0312 14:10:00.156384 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="extract-content" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.156405 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="extract-content" Mar 12 14:10:00 crc kubenswrapper[4778]: E0312 14:10:00.156439 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="registry-server" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.156447 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="registry-server" Mar 12 14:10:00 crc kubenswrapper[4778]: E0312 14:10:00.156461 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="extract-utilities" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.156469 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="extract-utilities" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.156735 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="6874a57b-b8de-4009-972e-c2c6d0635745" containerName="registry-server" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.157639 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.160983 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.162958 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.166039 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.168998 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555410-ptqps"] Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.309762 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4mtx\" (UniqueName: \"kubernetes.io/projected/3df6b9f3-72ae-4beb-b65c-c078aaf998ad-kube-api-access-s4mtx\") pod \"auto-csr-approver-29555410-ptqps\" (UID: \"3df6b9f3-72ae-4beb-b65c-c078aaf998ad\") " pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.411936 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4mtx\" (UniqueName: \"kubernetes.io/projected/3df6b9f3-72ae-4beb-b65c-c078aaf998ad-kube-api-access-s4mtx\") pod \"auto-csr-approver-29555410-ptqps\" (UID: \"3df6b9f3-72ae-4beb-b65c-c078aaf998ad\") " pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.437780 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4mtx\" (UniqueName: \"kubernetes.io/projected/3df6b9f3-72ae-4beb-b65c-c078aaf998ad-kube-api-access-s4mtx\") pod \"auto-csr-approver-29555410-ptqps\" (UID: \"3df6b9f3-72ae-4beb-b65c-c078aaf998ad\") " pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:00 crc kubenswrapper[4778]: I0312 14:10:00.489525 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:01 crc kubenswrapper[4778]: I0312 14:10:01.062996 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555410-ptqps"] Mar 12 14:10:01 crc kubenswrapper[4778]: I0312 14:10:01.187252 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555410-ptqps" event={"ID":"3df6b9f3-72ae-4beb-b65c-c078aaf998ad","Type":"ContainerStarted","Data":"30d6d0cde4fda952998df4697f22326d19f59d317ce023eebafb8133e5601af9"} Mar 12 14:10:03 crc kubenswrapper[4778]: I0312 14:10:03.206581 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555410-ptqps" event={"ID":"3df6b9f3-72ae-4beb-b65c-c078aaf998ad","Type":"ContainerStarted","Data":"fb40818927dd505564e4826e8d1f4316a9f1923eeaad7f19cc587698b0ad8339"} Mar 12 14:10:03 crc kubenswrapper[4778]: I0312 14:10:03.226052 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555410-ptqps" podStartSLOduration=1.574153421 podStartE2EDuration="3.226033959s" podCreationTimestamp="2026-03-12 14:10:00 +0000 UTC" firstStartedPulling="2026-03-12 14:10:01.065236263 +0000 UTC m=+3619.513931659" lastFinishedPulling="2026-03-12 14:10:02.717116801 +0000 UTC m=+3621.165812197" observedRunningTime="2026-03-12 14:10:03.219750561 +0000 UTC m=+3621.668445957" watchObservedRunningTime="2026-03-12 14:10:03.226033959 +0000 UTC m=+3621.674729355" Mar 12 14:10:04 crc kubenswrapper[4778]: I0312 14:10:04.217557 4778 generic.go:334] "Generic (PLEG): container finished" podID="3df6b9f3-72ae-4beb-b65c-c078aaf998ad" containerID="fb40818927dd505564e4826e8d1f4316a9f1923eeaad7f19cc587698b0ad8339" exitCode=0 Mar 12 14:10:04 crc kubenswrapper[4778]: I0312 14:10:04.217603 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555410-ptqps" event={"ID":"3df6b9f3-72ae-4beb-b65c-c078aaf998ad","Type":"ContainerDied","Data":"fb40818927dd505564e4826e8d1f4316a9f1923eeaad7f19cc587698b0ad8339"} Mar 12 14:10:05 crc kubenswrapper[4778]: I0312 14:10:05.837710 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:05 crc kubenswrapper[4778]: I0312 14:10:05.940397 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4mtx\" (UniqueName: \"kubernetes.io/projected/3df6b9f3-72ae-4beb-b65c-c078aaf998ad-kube-api-access-s4mtx\") pod \"3df6b9f3-72ae-4beb-b65c-c078aaf998ad\" (UID: \"3df6b9f3-72ae-4beb-b65c-c078aaf998ad\") " Mar 12 14:10:05 crc kubenswrapper[4778]: I0312 14:10:05.948023 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df6b9f3-72ae-4beb-b65c-c078aaf998ad-kube-api-access-s4mtx" (OuterVolumeSpecName: "kube-api-access-s4mtx") pod "3df6b9f3-72ae-4beb-b65c-c078aaf998ad" (UID: "3df6b9f3-72ae-4beb-b65c-c078aaf998ad"). InnerVolumeSpecName "kube-api-access-s4mtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:10:06 crc kubenswrapper[4778]: I0312 14:10:06.043467 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4mtx\" (UniqueName: \"kubernetes.io/projected/3df6b9f3-72ae-4beb-b65c-c078aaf998ad-kube-api-access-s4mtx\") on node \"crc\" DevicePath \"\"" Mar 12 14:10:06 crc kubenswrapper[4778]: I0312 14:10:06.237088 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555410-ptqps" event={"ID":"3df6b9f3-72ae-4beb-b65c-c078aaf998ad","Type":"ContainerDied","Data":"30d6d0cde4fda952998df4697f22326d19f59d317ce023eebafb8133e5601af9"} Mar 12 14:10:06 crc kubenswrapper[4778]: I0312 14:10:06.237140 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30d6d0cde4fda952998df4697f22326d19f59d317ce023eebafb8133e5601af9" Mar 12 14:10:06 crc kubenswrapper[4778]: I0312 14:10:06.237227 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555410-ptqps" Mar 12 14:10:06 crc kubenswrapper[4778]: I0312 14:10:06.368744 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555404-jwm56"] Mar 12 14:10:06 crc kubenswrapper[4778]: I0312 14:10:06.378511 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555404-jwm56"] Mar 12 14:10:08 crc kubenswrapper[4778]: I0312 14:10:08.267710 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67604e51-359f-4c7f-b7df-a4f215a87085" path="/var/lib/kubelet/pods/67604e51-359f-4c7f-b7df-a4f215a87085/volumes" Mar 12 14:10:20 crc kubenswrapper[4778]: I0312 14:10:20.823284 4778 scope.go:117] "RemoveContainer" containerID="67c3ac2c335344f6b2ef2e71132a310b8eda046527619858c13389f0ce08da63" Mar 12 14:10:28 crc kubenswrapper[4778]: I0312 14:10:28.558355 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:10:28 crc kubenswrapper[4778]: I0312 14:10:28.559779 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.557907 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.559552 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.559680 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.560552 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.560701 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" gracePeriod=600 Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.689512 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" exitCode=0 Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.689601 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2"} Mar 12 14:10:58 crc kubenswrapper[4778]: I0312 14:10:58.689915 4778 scope.go:117] "RemoveContainer" containerID="84eb4f64f5e57ea7581e624359f9a06ffee621fbf6407e2f32f007351966b81b" Mar 12 14:10:58 crc kubenswrapper[4778]: E0312 14:10:58.701214 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:10:59 crc kubenswrapper[4778]: I0312 14:10:59.703618 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:10:59 crc kubenswrapper[4778]: E0312 14:10:59.704410 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:11:12 crc kubenswrapper[4778]: I0312 14:11:12.261564 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:11:12 crc kubenswrapper[4778]: E0312 14:11:12.262442 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:11:27 crc kubenswrapper[4778]: I0312 14:11:27.253824 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:11:27 crc kubenswrapper[4778]: E0312 14:11:27.254684 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:11:42 crc kubenswrapper[4778]: I0312 14:11:42.259656 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:11:42 crc kubenswrapper[4778]: E0312 14:11:42.261649 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:11:54 crc kubenswrapper[4778]: I0312 14:11:54.254654 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:11:54 crc kubenswrapper[4778]: E0312 14:11:54.255257 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.151501 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555412-nvskv"] Mar 12 14:12:00 crc kubenswrapper[4778]: E0312 14:12:00.152403 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b9f3-72ae-4beb-b65c-c078aaf998ad" containerName="oc" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.152560 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b9f3-72ae-4beb-b65c-c078aaf998ad" containerName="oc" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.152851 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b9f3-72ae-4beb-b65c-c078aaf998ad" containerName="oc" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.153543 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.161235 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555412-nvskv"] Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.197891 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.197966 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.198268 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.251645 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqgsb\" (UniqueName: \"kubernetes.io/projected/f2a4f01e-04c1-43b0-8858-2d2334a828e5-kube-api-access-fqgsb\") pod \"auto-csr-approver-29555412-nvskv\" (UID: \"f2a4f01e-04c1-43b0-8858-2d2334a828e5\") " pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.354305 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqgsb\" (UniqueName: \"kubernetes.io/projected/f2a4f01e-04c1-43b0-8858-2d2334a828e5-kube-api-access-fqgsb\") pod \"auto-csr-approver-29555412-nvskv\" (UID: \"f2a4f01e-04c1-43b0-8858-2d2334a828e5\") " pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.384745 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqgsb\" (UniqueName: \"kubernetes.io/projected/f2a4f01e-04c1-43b0-8858-2d2334a828e5-kube-api-access-fqgsb\") pod \"auto-csr-approver-29555412-nvskv\" (UID: \"f2a4f01e-04c1-43b0-8858-2d2334a828e5\") " pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:00 crc kubenswrapper[4778]: I0312 14:12:00.529376 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:01 crc kubenswrapper[4778]: I0312 14:12:01.041115 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555412-nvskv"] Mar 12 14:12:01 crc kubenswrapper[4778]: I0312 14:12:01.946142 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555412-nvskv" event={"ID":"f2a4f01e-04c1-43b0-8858-2d2334a828e5","Type":"ContainerStarted","Data":"f1b77ae6764356f91826bfb42ac02b4ab68a819c21ed01bf516b19b118d59157"} Mar 12 14:12:02 crc kubenswrapper[4778]: I0312 14:12:02.955171 4778 generic.go:334] "Generic (PLEG): container finished" podID="f2a4f01e-04c1-43b0-8858-2d2334a828e5" containerID="a0fce8e55d131bb482515dc65d16783265a86aa260db0e75ebb8541d77da26bd" exitCode=0 Mar 12 14:12:02 crc kubenswrapper[4778]: I0312 14:12:02.955340 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555412-nvskv" event={"ID":"f2a4f01e-04c1-43b0-8858-2d2334a828e5","Type":"ContainerDied","Data":"a0fce8e55d131bb482515dc65d16783265a86aa260db0e75ebb8541d77da26bd"} Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.482545 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.636656 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqgsb\" (UniqueName: \"kubernetes.io/projected/f2a4f01e-04c1-43b0-8858-2d2334a828e5-kube-api-access-fqgsb\") pod \"f2a4f01e-04c1-43b0-8858-2d2334a828e5\" (UID: \"f2a4f01e-04c1-43b0-8858-2d2334a828e5\") " Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.643974 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2a4f01e-04c1-43b0-8858-2d2334a828e5-kube-api-access-fqgsb" (OuterVolumeSpecName: "kube-api-access-fqgsb") pod "f2a4f01e-04c1-43b0-8858-2d2334a828e5" (UID: "f2a4f01e-04c1-43b0-8858-2d2334a828e5"). InnerVolumeSpecName "kube-api-access-fqgsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.739708 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqgsb\" (UniqueName: \"kubernetes.io/projected/f2a4f01e-04c1-43b0-8858-2d2334a828e5-kube-api-access-fqgsb\") on node \"crc\" DevicePath \"\"" Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.974305 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555412-nvskv" event={"ID":"f2a4f01e-04c1-43b0-8858-2d2334a828e5","Type":"ContainerDied","Data":"f1b77ae6764356f91826bfb42ac02b4ab68a819c21ed01bf516b19b118d59157"} Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.974345 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b77ae6764356f91826bfb42ac02b4ab68a819c21ed01bf516b19b118d59157" Mar 12 14:12:04 crc kubenswrapper[4778]: I0312 14:12:04.974353 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555412-nvskv" Mar 12 14:12:05 crc kubenswrapper[4778]: I0312 14:12:05.555050 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555406-44v2c"] Mar 12 14:12:05 crc kubenswrapper[4778]: I0312 14:12:05.567167 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555406-44v2c"] Mar 12 14:12:06 crc kubenswrapper[4778]: I0312 14:12:06.263873 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e7c143c-4173-450a-afa1-587a3927f2d4" path="/var/lib/kubelet/pods/2e7c143c-4173-450a-afa1-587a3927f2d4/volumes" Mar 12 14:12:07 crc kubenswrapper[4778]: I0312 14:12:07.254158 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:12:07 crc kubenswrapper[4778]: E0312 14:12:07.255027 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:12:20 crc kubenswrapper[4778]: I0312 14:12:20.924355 4778 scope.go:117] "RemoveContainer" containerID="2c9bf5717fd9b2c8602b788cdc193d4c283ef18c6a74310ea29b1e044df19e27" Mar 12 14:12:22 crc kubenswrapper[4778]: I0312 14:12:22.260262 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:12:22 crc kubenswrapper[4778]: E0312 14:12:22.260904 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:12:36 crc kubenswrapper[4778]: I0312 14:12:36.254047 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:12:36 crc kubenswrapper[4778]: E0312 14:12:36.254834 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:12:48 crc kubenswrapper[4778]: I0312 14:12:48.255728 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:12:48 crc kubenswrapper[4778]: E0312 14:12:48.256550 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:13:01 crc kubenswrapper[4778]: I0312 14:13:01.253650 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:13:01 crc kubenswrapper[4778]: E0312 14:13:01.254709 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:13:15 crc kubenswrapper[4778]: I0312 14:13:15.254700 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:13:15 crc kubenswrapper[4778]: E0312 14:13:15.255465 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:13:29 crc kubenswrapper[4778]: I0312 14:13:29.321964 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:13:29 crc kubenswrapper[4778]: E0312 14:13:29.322855 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:13:40 crc kubenswrapper[4778]: I0312 14:13:40.254382 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:13:40 crc kubenswrapper[4778]: E0312 14:13:40.255113 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:13:55 crc kubenswrapper[4778]: I0312 14:13:55.254679 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:13:55 crc kubenswrapper[4778]: E0312 14:13:55.255558 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.142210 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555414-sk27k"] Mar 12 14:14:00 crc kubenswrapper[4778]: E0312 14:14:00.144339 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a4f01e-04c1-43b0-8858-2d2334a828e5" containerName="oc" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.144453 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a4f01e-04c1-43b0-8858-2d2334a828e5" containerName="oc" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.144772 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a4f01e-04c1-43b0-8858-2d2334a828e5" containerName="oc" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.145576 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.147736 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.147736 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.148727 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.166849 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555414-sk27k"] Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.199151 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfcb9\" (UniqueName: \"kubernetes.io/projected/066c6c88-5aea-4678-88a0-ec5c556ee008-kube-api-access-rfcb9\") pod \"auto-csr-approver-29555414-sk27k\" (UID: \"066c6c88-5aea-4678-88a0-ec5c556ee008\") " pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.302609 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfcb9\" (UniqueName: \"kubernetes.io/projected/066c6c88-5aea-4678-88a0-ec5c556ee008-kube-api-access-rfcb9\") pod \"auto-csr-approver-29555414-sk27k\" (UID: \"066c6c88-5aea-4678-88a0-ec5c556ee008\") " pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.325174 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfcb9\" (UniqueName: \"kubernetes.io/projected/066c6c88-5aea-4678-88a0-ec5c556ee008-kube-api-access-rfcb9\") pod \"auto-csr-approver-29555414-sk27k\" (UID: \"066c6c88-5aea-4678-88a0-ec5c556ee008\") " pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.499301 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:00 crc kubenswrapper[4778]: I0312 14:14:00.994414 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555414-sk27k"] Mar 12 14:14:01 crc kubenswrapper[4778]: I0312 14:14:01.009578 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:14:01 crc kubenswrapper[4778]: I0312 14:14:01.977846 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555414-sk27k" event={"ID":"066c6c88-5aea-4678-88a0-ec5c556ee008","Type":"ContainerStarted","Data":"25dbd3dcdc293d45446a6218fa50871970a46f1bd6b9c3d4c1547536ec32ea1c"} Mar 12 14:14:02 crc kubenswrapper[4778]: I0312 14:14:02.998535 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555414-sk27k" event={"ID":"066c6c88-5aea-4678-88a0-ec5c556ee008","Type":"ContainerStarted","Data":"3925769f3c54add574a18597a06eea490ae5d1cab077561f5ae8b471c0db5519"} Mar 12 14:14:04 crc kubenswrapper[4778]: I0312 14:14:04.014786 4778 generic.go:334] "Generic (PLEG): container finished" podID="066c6c88-5aea-4678-88a0-ec5c556ee008" containerID="3925769f3c54add574a18597a06eea490ae5d1cab077561f5ae8b471c0db5519" exitCode=0 Mar 12 14:14:04 crc kubenswrapper[4778]: I0312 14:14:04.014993 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555414-sk27k" event={"ID":"066c6c88-5aea-4678-88a0-ec5c556ee008","Type":"ContainerDied","Data":"3925769f3c54add574a18597a06eea490ae5d1cab077561f5ae8b471c0db5519"} Mar 12 14:14:05 crc kubenswrapper[4778]: I0312 14:14:05.574264 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:05 crc kubenswrapper[4778]: I0312 14:14:05.707548 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfcb9\" (UniqueName: \"kubernetes.io/projected/066c6c88-5aea-4678-88a0-ec5c556ee008-kube-api-access-rfcb9\") pod \"066c6c88-5aea-4678-88a0-ec5c556ee008\" (UID: \"066c6c88-5aea-4678-88a0-ec5c556ee008\") " Mar 12 14:14:05 crc kubenswrapper[4778]: I0312 14:14:05.717808 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/066c6c88-5aea-4678-88a0-ec5c556ee008-kube-api-access-rfcb9" (OuterVolumeSpecName: "kube-api-access-rfcb9") pod "066c6c88-5aea-4678-88a0-ec5c556ee008" (UID: "066c6c88-5aea-4678-88a0-ec5c556ee008"). InnerVolumeSpecName "kube-api-access-rfcb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:14:05 crc kubenswrapper[4778]: I0312 14:14:05.810478 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfcb9\" (UniqueName: \"kubernetes.io/projected/066c6c88-5aea-4678-88a0-ec5c556ee008-kube-api-access-rfcb9\") on node \"crc\" DevicePath \"\"" Mar 12 14:14:06 crc kubenswrapper[4778]: I0312 14:14:06.033431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555414-sk27k" event={"ID":"066c6c88-5aea-4678-88a0-ec5c556ee008","Type":"ContainerDied","Data":"25dbd3dcdc293d45446a6218fa50871970a46f1bd6b9c3d4c1547536ec32ea1c"} Mar 12 14:14:06 crc kubenswrapper[4778]: I0312 14:14:06.033711 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25dbd3dcdc293d45446a6218fa50871970a46f1bd6b9c3d4c1547536ec32ea1c" Mar 12 14:14:06 crc kubenswrapper[4778]: I0312 14:14:06.033485 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555414-sk27k" Mar 12 14:14:06 crc kubenswrapper[4778]: I0312 14:14:06.650093 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555408-92kzm"] Mar 12 14:14:06 crc kubenswrapper[4778]: I0312 14:14:06.661524 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555408-92kzm"] Mar 12 14:14:07 crc kubenswrapper[4778]: I0312 14:14:07.253679 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:14:07 crc kubenswrapper[4778]: E0312 14:14:07.254014 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:14:08 crc kubenswrapper[4778]: I0312 14:14:08.268359 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2" path="/var/lib/kubelet/pods/f7d6c255-3117-4dbe-b3d6-23f7be9f1cf2/volumes" Mar 12 14:14:21 crc kubenswrapper[4778]: I0312 14:14:21.018527 4778 scope.go:117] "RemoveContainer" containerID="9557c198d563e8bc3c1bd0c3db7f0caaf03fda89ffb3294270c9a1e6bbdd5486" Mar 12 14:14:21 crc kubenswrapper[4778]: I0312 14:14:21.253865 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:14:21 crc kubenswrapper[4778]: E0312 14:14:21.254395 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:14:36 crc kubenswrapper[4778]: I0312 14:14:36.254143 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:14:36 crc kubenswrapper[4778]: E0312 14:14:36.255083 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:14:51 crc kubenswrapper[4778]: I0312 14:14:51.253439 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:14:51 crc kubenswrapper[4778]: E0312 14:14:51.253855 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.153156 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r"] Mar 12 14:15:00 crc kubenswrapper[4778]: E0312 14:15:00.154327 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="066c6c88-5aea-4678-88a0-ec5c556ee008" containerName="oc" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.154343 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="066c6c88-5aea-4678-88a0-ec5c556ee008" containerName="oc" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.154622 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="066c6c88-5aea-4678-88a0-ec5c556ee008" containerName="oc" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.155478 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.158615 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.160655 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.161682 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r"] Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.324676 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-config-volume\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.325105 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-secret-volume\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.325153 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d7cx\" (UniqueName: \"kubernetes.io/projected/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-kube-api-access-8d7cx\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.427091 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-secret-volume\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.427176 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d7cx\" (UniqueName: \"kubernetes.io/projected/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-kube-api-access-8d7cx\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.427265 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-config-volume\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.428362 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-config-volume\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.433130 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-secret-volume\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.447068 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d7cx\" (UniqueName: \"kubernetes.io/projected/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-kube-api-access-8d7cx\") pod \"collect-profiles-29555415-jjk6r\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.484317 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:00 crc kubenswrapper[4778]: I0312 14:15:00.980805 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r"] Mar 12 14:15:01 crc kubenswrapper[4778]: I0312 14:15:01.540988 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" event={"ID":"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc","Type":"ContainerStarted","Data":"e47d44b34f9f52eb0c1249aedb361a64e96dcc50294b7036054124a9fc860b25"} Mar 12 14:15:01 crc kubenswrapper[4778]: I0312 14:15:01.541235 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" event={"ID":"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc","Type":"ContainerStarted","Data":"e2b3b34963b216b0a4e58389ce9a9953591924a46c7f59a6c519e692e6e3e738"} Mar 12 14:15:01 crc kubenswrapper[4778]: I0312 14:15:01.565569 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" podStartSLOduration=1.565544987 podStartE2EDuration="1.565544987s" podCreationTimestamp="2026-03-12 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 14:15:01.563697905 +0000 UTC m=+3920.012393291" watchObservedRunningTime="2026-03-12 14:15:01.565544987 +0000 UTC m=+3920.014240393" Mar 12 14:15:02 crc kubenswrapper[4778]: I0312 14:15:02.549671 4778 generic.go:334] "Generic (PLEG): container finished" podID="9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" containerID="e47d44b34f9f52eb0c1249aedb361a64e96dcc50294b7036054124a9fc860b25" exitCode=0 Mar 12 14:15:02 crc kubenswrapper[4778]: I0312 14:15:02.549714 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" event={"ID":"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc","Type":"ContainerDied","Data":"e47d44b34f9f52eb0c1249aedb361a64e96dcc50294b7036054124a9fc860b25"} Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.125859 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.203243 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-config-volume\") pod \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.203315 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d7cx\" (UniqueName: \"kubernetes.io/projected/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-kube-api-access-8d7cx\") pod \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.203672 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-config-volume" (OuterVolumeSpecName: "config-volume") pod "9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" (UID: "9c6027ea-d1ed-4df0-bbe7-6904d2722fbc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.204285 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-secret-volume\") pod \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\" (UID: \"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc\") " Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.205047 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.208317 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-kube-api-access-8d7cx" (OuterVolumeSpecName: "kube-api-access-8d7cx") pod "9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" (UID: "9c6027ea-d1ed-4df0-bbe7-6904d2722fbc"). InnerVolumeSpecName "kube-api-access-8d7cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.208664 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" (UID: "9c6027ea-d1ed-4df0-bbe7-6904d2722fbc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.259655 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:15:04 crc kubenswrapper[4778]: E0312 14:15:04.259865 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.306906 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.306949 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d7cx\" (UniqueName: \"kubernetes.io/projected/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc-kube-api-access-8d7cx\") on node \"crc\" DevicePath \"\"" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.575457 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" event={"ID":"9c6027ea-d1ed-4df0-bbe7-6904d2722fbc","Type":"ContainerDied","Data":"e2b3b34963b216b0a4e58389ce9a9953591924a46c7f59a6c519e692e6e3e738"} Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.576087 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2b3b34963b216b0a4e58389ce9a9953591924a46c7f59a6c519e692e6e3e738" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.576261 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r" Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.642296 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f"] Mar 12 14:15:04 crc kubenswrapper[4778]: I0312 14:15:04.663456 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555370-zcp5f"] Mar 12 14:15:06 crc kubenswrapper[4778]: I0312 14:15:06.266819 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf03685-d980-41f0-bbc5-84b9ae0ce1df" path="/var/lib/kubelet/pods/8bf03685-d980-41f0-bbc5-84b9ae0ce1df/volumes" Mar 12 14:15:15 crc kubenswrapper[4778]: I0312 14:15:15.254359 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:15:15 crc kubenswrapper[4778]: E0312 14:15:15.255240 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:15:21 crc kubenswrapper[4778]: I0312 14:15:21.090068 4778 scope.go:117] "RemoveContainer" containerID="fa067a709ad1af5d5b9327929891ffc04839dd2d8aba3cc70c48dbfeabd353b9" Mar 12 14:15:28 crc kubenswrapper[4778]: I0312 14:15:28.258535 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:15:28 crc kubenswrapper[4778]: E0312 14:15:28.259436 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.144413 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x7hrm"] Mar 12 14:15:40 crc kubenswrapper[4778]: E0312 14:15:40.145502 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" containerName="collect-profiles" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.145517 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" containerName="collect-profiles" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.145766 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" containerName="collect-profiles" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.147461 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.158407 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x7hrm"] Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.265637 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b94cm\" (UniqueName: \"kubernetes.io/projected/cc5d1f36-7c1d-4e41-8b22-e332bc157137-kube-api-access-b94cm\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.265696 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-utilities\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.265857 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-catalog-content\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.367930 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-catalog-content\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.368054 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b94cm\" (UniqueName: \"kubernetes.io/projected/cc5d1f36-7c1d-4e41-8b22-e332bc157137-kube-api-access-b94cm\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.368086 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-utilities\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.368563 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-utilities\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.369125 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-catalog-content\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.387743 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b94cm\" (UniqueName: \"kubernetes.io/projected/cc5d1f36-7c1d-4e41-8b22-e332bc157137-kube-api-access-b94cm\") pod \"certified-operators-x7hrm\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:40 crc kubenswrapper[4778]: I0312 14:15:40.470194 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:41 crc kubenswrapper[4778]: I0312 14:15:41.004038 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x7hrm"] Mar 12 14:15:41 crc kubenswrapper[4778]: I0312 14:15:41.944734 4778 generic.go:334] "Generic (PLEG): container finished" podID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerID="64f499cf2cc148115dc0a26243c38d7001d6394769d13ddee3ac1cad8976c318" exitCode=0 Mar 12 14:15:41 crc kubenswrapper[4778]: I0312 14:15:41.945061 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerDied","Data":"64f499cf2cc148115dc0a26243c38d7001d6394769d13ddee3ac1cad8976c318"} Mar 12 14:15:41 crc kubenswrapper[4778]: I0312 14:15:41.945095 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerStarted","Data":"26c7df7cd30e2c7b7c46b23a72c15208ba43c942293bb8165b5463087a90fd16"} Mar 12 14:15:43 crc kubenswrapper[4778]: I0312 14:15:43.254510 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:15:43 crc kubenswrapper[4778]: E0312 14:15:43.255331 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:15:43 crc kubenswrapper[4778]: I0312 14:15:43.963127 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerStarted","Data":"1c4ad35649138a441fab3a7d2a2c15667596a5cea9441d67b18d6bf6786027d9"} Mar 12 14:15:45 crc kubenswrapper[4778]: I0312 14:15:45.987942 4778 generic.go:334] "Generic (PLEG): container finished" podID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerID="1c4ad35649138a441fab3a7d2a2c15667596a5cea9441d67b18d6bf6786027d9" exitCode=0 Mar 12 14:15:45 crc kubenswrapper[4778]: I0312 14:15:45.988029 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerDied","Data":"1c4ad35649138a441fab3a7d2a2c15667596a5cea9441d67b18d6bf6786027d9"} Mar 12 14:15:47 crc kubenswrapper[4778]: I0312 14:15:47.003014 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerStarted","Data":"6ccc0779d6618322241261d3f4156d21061888679b769f8c1765e9abe88eab3e"} Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.471343 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.471872 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.526223 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.550265 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x7hrm" podStartSLOduration=5.831476572 podStartE2EDuration="10.550250597s" podCreationTimestamp="2026-03-12 14:15:40 +0000 UTC" firstStartedPulling="2026-03-12 14:15:41.947305323 +0000 UTC m=+3960.396000709" lastFinishedPulling="2026-03-12 14:15:46.666079328 +0000 UTC m=+3965.114774734" observedRunningTime="2026-03-12 14:15:47.032154091 +0000 UTC m=+3965.480849487" watchObservedRunningTime="2026-03-12 14:15:50.550250597 +0000 UTC m=+3968.998945993" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.773359 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nhcwl"] Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.775637 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.794308 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nhcwl"] Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.869544 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-catalog-content\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.869717 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrvq2\" (UniqueName: \"kubernetes.io/projected/7971c2f5-a365-405f-9acf-0ef296dcedcf-kube-api-access-hrvq2\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.869781 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-utilities\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.971456 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrvq2\" (UniqueName: \"kubernetes.io/projected/7971c2f5-a365-405f-9acf-0ef296dcedcf-kube-api-access-hrvq2\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.971546 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-utilities\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.971683 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-catalog-content\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.972142 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-utilities\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.972423 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-catalog-content\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:50 crc kubenswrapper[4778]: I0312 14:15:50.998300 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrvq2\" (UniqueName: \"kubernetes.io/projected/7971c2f5-a365-405f-9acf-0ef296dcedcf-kube-api-access-hrvq2\") pod \"community-operators-nhcwl\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:51 crc kubenswrapper[4778]: I0312 14:15:51.091440 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:51 crc kubenswrapper[4778]: I0312 14:15:51.094693 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:15:51 crc kubenswrapper[4778]: I0312 14:15:51.656129 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nhcwl"] Mar 12 14:15:52 crc kubenswrapper[4778]: I0312 14:15:52.052115 4778 generic.go:334] "Generic (PLEG): container finished" podID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerID="4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494" exitCode=0 Mar 12 14:15:52 crc kubenswrapper[4778]: I0312 14:15:52.052285 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerDied","Data":"4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494"} Mar 12 14:15:52 crc kubenswrapper[4778]: I0312 14:15:52.052530 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerStarted","Data":"332bc25c0b47d707315b376f0305bf18f3b8ef2e6fc97b50c01fdda78dbf3fe1"} Mar 12 14:15:53 crc kubenswrapper[4778]: I0312 14:15:53.360826 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x7hrm"] Mar 12 14:15:53 crc kubenswrapper[4778]: I0312 14:15:53.361270 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x7hrm" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="registry-server" containerID="cri-o://6ccc0779d6618322241261d3f4156d21061888679b769f8c1765e9abe88eab3e" gracePeriod=2 Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.075140 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerStarted","Data":"96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1"} Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.078436 4778 generic.go:334] "Generic (PLEG): container finished" podID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerID="6ccc0779d6618322241261d3f4156d21061888679b769f8c1765e9abe88eab3e" exitCode=0 Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.078607 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerDied","Data":"6ccc0779d6618322241261d3f4156d21061888679b769f8c1765e9abe88eab3e"} Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.253507 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:15:54 crc kubenswrapper[4778]: E0312 14:15:54.253751 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.277312 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.445969 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b94cm\" (UniqueName: \"kubernetes.io/projected/cc5d1f36-7c1d-4e41-8b22-e332bc157137-kube-api-access-b94cm\") pod \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.446041 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-utilities\") pod \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.446095 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-catalog-content\") pod \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\" (UID: \"cc5d1f36-7c1d-4e41-8b22-e332bc157137\") " Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.447667 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-utilities" (OuterVolumeSpecName: "utilities") pod "cc5d1f36-7c1d-4e41-8b22-e332bc157137" (UID: "cc5d1f36-7c1d-4e41-8b22-e332bc157137"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.455668 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc5d1f36-7c1d-4e41-8b22-e332bc157137-kube-api-access-b94cm" (OuterVolumeSpecName: "kube-api-access-b94cm") pod "cc5d1f36-7c1d-4e41-8b22-e332bc157137" (UID: "cc5d1f36-7c1d-4e41-8b22-e332bc157137"). InnerVolumeSpecName "kube-api-access-b94cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.519482 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc5d1f36-7c1d-4e41-8b22-e332bc157137" (UID: "cc5d1f36-7c1d-4e41-8b22-e332bc157137"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.548111 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b94cm\" (UniqueName: \"kubernetes.io/projected/cc5d1f36-7c1d-4e41-8b22-e332bc157137-kube-api-access-b94cm\") on node \"crc\" DevicePath \"\"" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.548147 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:15:54 crc kubenswrapper[4778]: I0312 14:15:54.548157 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc5d1f36-7c1d-4e41-8b22-e332bc157137-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.090888 4778 generic.go:334] "Generic (PLEG): container finished" podID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerID="96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1" exitCode=0 Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.091360 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerDied","Data":"96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1"} Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.094204 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7hrm" event={"ID":"cc5d1f36-7c1d-4e41-8b22-e332bc157137","Type":"ContainerDied","Data":"26c7df7cd30e2c7b7c46b23a72c15208ba43c942293bb8165b5463087a90fd16"} Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.094251 4778 scope.go:117] "RemoveContainer" containerID="6ccc0779d6618322241261d3f4156d21061888679b769f8c1765e9abe88eab3e" Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.094266 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7hrm" Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.119856 4778 scope.go:117] "RemoveContainer" containerID="1c4ad35649138a441fab3a7d2a2c15667596a5cea9441d67b18d6bf6786027d9" Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.154931 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x7hrm"] Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.175373 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x7hrm"] Mar 12 14:15:55 crc kubenswrapper[4778]: I0312 14:15:55.175880 4778 scope.go:117] "RemoveContainer" containerID="64f499cf2cc148115dc0a26243c38d7001d6394769d13ddee3ac1cad8976c318" Mar 12 14:15:56 crc kubenswrapper[4778]: I0312 14:15:56.107940 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerStarted","Data":"33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93"} Mar 12 14:15:56 crc kubenswrapper[4778]: I0312 14:15:56.129585 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nhcwl" podStartSLOduration=2.666888212 podStartE2EDuration="6.129565784s" podCreationTimestamp="2026-03-12 14:15:50 +0000 UTC" firstStartedPulling="2026-03-12 14:15:52.054738887 +0000 UTC m=+3970.503434283" lastFinishedPulling="2026-03-12 14:15:55.517416469 +0000 UTC m=+3973.966111855" observedRunningTime="2026-03-12 14:15:56.126173358 +0000 UTC m=+3974.574868764" watchObservedRunningTime="2026-03-12 14:15:56.129565784 +0000 UTC m=+3974.578261200" Mar 12 14:15:56 crc kubenswrapper[4778]: I0312 14:15:56.265396 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" path="/var/lib/kubelet/pods/cc5d1f36-7c1d-4e41-8b22-e332bc157137/volumes" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.161498 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555416-qx4gr"] Mar 12 14:16:00 crc kubenswrapper[4778]: E0312 14:16:00.162659 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="extract-utilities" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.162679 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="extract-utilities" Mar 12 14:16:00 crc kubenswrapper[4778]: E0312 14:16:00.162712 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="registry-server" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.162722 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="registry-server" Mar 12 14:16:00 crc kubenswrapper[4778]: E0312 14:16:00.162739 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="extract-content" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.162748 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="extract-content" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.162994 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc5d1f36-7c1d-4e41-8b22-e332bc157137" containerName="registry-server" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.163848 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.166544 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.167251 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.167263 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.172734 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555416-qx4gr"] Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.261463 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg594\" (UniqueName: \"kubernetes.io/projected/43680ad6-62d2-4e00-a38b-e87d712af9a7-kube-api-access-wg594\") pod \"auto-csr-approver-29555416-qx4gr\" (UID: \"43680ad6-62d2-4e00-a38b-e87d712af9a7\") " pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.362726 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg594\" (UniqueName: \"kubernetes.io/projected/43680ad6-62d2-4e00-a38b-e87d712af9a7-kube-api-access-wg594\") pod \"auto-csr-approver-29555416-qx4gr\" (UID: \"43680ad6-62d2-4e00-a38b-e87d712af9a7\") " pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.385425 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg594\" (UniqueName: \"kubernetes.io/projected/43680ad6-62d2-4e00-a38b-e87d712af9a7-kube-api-access-wg594\") pod \"auto-csr-approver-29555416-qx4gr\" (UID: \"43680ad6-62d2-4e00-a38b-e87d712af9a7\") " pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:00 crc kubenswrapper[4778]: I0312 14:16:00.500873 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.002080 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555416-qx4gr"] Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.095155 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.095237 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.146113 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.172360 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" event={"ID":"43680ad6-62d2-4e00-a38b-e87d712af9a7","Type":"ContainerStarted","Data":"6e1d6cb507d001f6a763ab8ac3b873837b5430cdf2bccfa5deae43ad5b72ed5d"} Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.223445 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:16:01 crc kubenswrapper[4778]: I0312 14:16:01.379921 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nhcwl"] Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.191251 4778 generic.go:334] "Generic (PLEG): container finished" podID="43680ad6-62d2-4e00-a38b-e87d712af9a7" containerID="95522e18d302b349263bfd01c0f317ec6a698231318f520fcd1ed51d7aa504cc" exitCode=0 Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.191669 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nhcwl" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="registry-server" containerID="cri-o://33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93" gracePeriod=2 Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.191995 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" event={"ID":"43680ad6-62d2-4e00-a38b-e87d712af9a7","Type":"ContainerDied","Data":"95522e18d302b349263bfd01c0f317ec6a698231318f520fcd1ed51d7aa504cc"} Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.833882 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.944385 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-utilities\") pod \"7971c2f5-a365-405f-9acf-0ef296dcedcf\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.944578 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrvq2\" (UniqueName: \"kubernetes.io/projected/7971c2f5-a365-405f-9acf-0ef296dcedcf-kube-api-access-hrvq2\") pod \"7971c2f5-a365-405f-9acf-0ef296dcedcf\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.944688 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-catalog-content\") pod \"7971c2f5-a365-405f-9acf-0ef296dcedcf\" (UID: \"7971c2f5-a365-405f-9acf-0ef296dcedcf\") " Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.945269 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-utilities" (OuterVolumeSpecName: "utilities") pod "7971c2f5-a365-405f-9acf-0ef296dcedcf" (UID: "7971c2f5-a365-405f-9acf-0ef296dcedcf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:16:03 crc kubenswrapper[4778]: I0312 14:16:03.952704 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7971c2f5-a365-405f-9acf-0ef296dcedcf-kube-api-access-hrvq2" (OuterVolumeSpecName: "kube-api-access-hrvq2") pod "7971c2f5-a365-405f-9acf-0ef296dcedcf" (UID: "7971c2f5-a365-405f-9acf-0ef296dcedcf"). InnerVolumeSpecName "kube-api-access-hrvq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.012564 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7971c2f5-a365-405f-9acf-0ef296dcedcf" (UID: "7971c2f5-a365-405f-9acf-0ef296dcedcf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.046730 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.046772 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7971c2f5-a365-405f-9acf-0ef296dcedcf-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.046783 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrvq2\" (UniqueName: \"kubernetes.io/projected/7971c2f5-a365-405f-9acf-0ef296dcedcf-kube-api-access-hrvq2\") on node \"crc\" DevicePath \"\"" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.203441 4778 generic.go:334] "Generic (PLEG): container finished" podID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerID="33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93" exitCode=0 Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.203518 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nhcwl" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.203518 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerDied","Data":"33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93"} Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.203920 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nhcwl" event={"ID":"7971c2f5-a365-405f-9acf-0ef296dcedcf","Type":"ContainerDied","Data":"332bc25c0b47d707315b376f0305bf18f3b8ef2e6fc97b50c01fdda78dbf3fe1"} Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.203950 4778 scope.go:117] "RemoveContainer" containerID="33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.278875 4778 scope.go:117] "RemoveContainer" containerID="96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.285821 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nhcwl"] Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.285855 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nhcwl"] Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.313468 4778 scope.go:117] "RemoveContainer" containerID="4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.355274 4778 scope.go:117] "RemoveContainer" containerID="33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93" Mar 12 14:16:04 crc kubenswrapper[4778]: E0312 14:16:04.360425 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93\": container with ID starting with 33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93 not found: ID does not exist" containerID="33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.360480 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93"} err="failed to get container status \"33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93\": rpc error: code = NotFound desc = could not find container \"33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93\": container with ID starting with 33f839b589bdd36727509b18dd4d904b894fa2e2c1aae95639ee07888bcf2c93 not found: ID does not exist" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.360510 4778 scope.go:117] "RemoveContainer" containerID="96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1" Mar 12 14:16:04 crc kubenswrapper[4778]: E0312 14:16:04.361126 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1\": container with ID starting with 96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1 not found: ID does not exist" containerID="96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.361177 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1"} err="failed to get container status \"96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1\": rpc error: code = NotFound desc = could not find container \"96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1\": container with ID starting with 96aade208371a8c6296565d119296613b2c7f130c7ab3677910649b35169fdf1 not found: ID does not exist" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.361227 4778 scope.go:117] "RemoveContainer" containerID="4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494" Mar 12 14:16:04 crc kubenswrapper[4778]: E0312 14:16:04.361665 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494\": container with ID starting with 4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494 not found: ID does not exist" containerID="4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.361694 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494"} err="failed to get container status \"4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494\": rpc error: code = NotFound desc = could not find container \"4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494\": container with ID starting with 4736be3491cb2ad9c4a9ad132c1d7d595ed12aef66987357666acd2528e82494 not found: ID does not exist" Mar 12 14:16:04 crc kubenswrapper[4778]: I0312 14:16:04.866368 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.068807 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg594\" (UniqueName: \"kubernetes.io/projected/43680ad6-62d2-4e00-a38b-e87d712af9a7-kube-api-access-wg594\") pod \"43680ad6-62d2-4e00-a38b-e87d712af9a7\" (UID: \"43680ad6-62d2-4e00-a38b-e87d712af9a7\") " Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.073842 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43680ad6-62d2-4e00-a38b-e87d712af9a7-kube-api-access-wg594" (OuterVolumeSpecName: "kube-api-access-wg594") pod "43680ad6-62d2-4e00-a38b-e87d712af9a7" (UID: "43680ad6-62d2-4e00-a38b-e87d712af9a7"). InnerVolumeSpecName "kube-api-access-wg594". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.170983 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg594\" (UniqueName: \"kubernetes.io/projected/43680ad6-62d2-4e00-a38b-e87d712af9a7-kube-api-access-wg594\") on node \"crc\" DevicePath \"\"" Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.214361 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" event={"ID":"43680ad6-62d2-4e00-a38b-e87d712af9a7","Type":"ContainerDied","Data":"6e1d6cb507d001f6a763ab8ac3b873837b5430cdf2bccfa5deae43ad5b72ed5d"} Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.214392 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555416-qx4gr" Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.214414 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e1d6cb507d001f6a763ab8ac3b873837b5430cdf2bccfa5deae43ad5b72ed5d" Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.959620 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555410-ptqps"] Mar 12 14:16:05 crc kubenswrapper[4778]: I0312 14:16:05.968902 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555410-ptqps"] Mar 12 14:16:06 crc kubenswrapper[4778]: I0312 14:16:06.265623 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df6b9f3-72ae-4beb-b65c-c078aaf998ad" path="/var/lib/kubelet/pods/3df6b9f3-72ae-4beb-b65c-c078aaf998ad/volumes" Mar 12 14:16:06 crc kubenswrapper[4778]: I0312 14:16:06.266520 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" path="/var/lib/kubelet/pods/7971c2f5-a365-405f-9acf-0ef296dcedcf/volumes" Mar 12 14:16:08 crc kubenswrapper[4778]: I0312 14:16:08.259417 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:16:09 crc kubenswrapper[4778]: I0312 14:16:09.253807 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"a6494f4559bd62f54e1e656b9c39bd8218ddcffc6f2d4766fd788af23c632a2c"} Mar 12 14:16:21 crc kubenswrapper[4778]: I0312 14:16:21.146928 4778 scope.go:117] "RemoveContainer" containerID="fb40818927dd505564e4826e8d1f4316a9f1923eeaad7f19cc587698b0ad8339" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.151838 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555418-xsl6h"] Mar 12 14:18:00 crc kubenswrapper[4778]: E0312 14:18:00.152954 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="registry-server" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.152973 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="registry-server" Mar 12 14:18:00 crc kubenswrapper[4778]: E0312 14:18:00.153012 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="extract-content" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.153020 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="extract-content" Mar 12 14:18:00 crc kubenswrapper[4778]: E0312 14:18:00.153046 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="extract-utilities" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.153054 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="extract-utilities" Mar 12 14:18:00 crc kubenswrapper[4778]: E0312 14:18:00.153067 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43680ad6-62d2-4e00-a38b-e87d712af9a7" containerName="oc" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.153074 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="43680ad6-62d2-4e00-a38b-e87d712af9a7" containerName="oc" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.153308 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7971c2f5-a365-405f-9acf-0ef296dcedcf" containerName="registry-server" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.153332 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="43680ad6-62d2-4e00-a38b-e87d712af9a7" containerName="oc" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.154175 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.156758 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.157019 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.157224 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.173013 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555418-xsl6h"] Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.265494 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd9r9\" (UniqueName: \"kubernetes.io/projected/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9-kube-api-access-xd9r9\") pod \"auto-csr-approver-29555418-xsl6h\" (UID: \"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9\") " pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.367814 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd9r9\" (UniqueName: \"kubernetes.io/projected/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9-kube-api-access-xd9r9\") pod \"auto-csr-approver-29555418-xsl6h\" (UID: \"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9\") " pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.394114 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd9r9\" (UniqueName: \"kubernetes.io/projected/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9-kube-api-access-xd9r9\") pod \"auto-csr-approver-29555418-xsl6h\" (UID: \"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9\") " pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.496409 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:00 crc kubenswrapper[4778]: I0312 14:18:00.977156 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555418-xsl6h"] Mar 12 14:18:01 crc kubenswrapper[4778]: I0312 14:18:01.378198 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" event={"ID":"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9","Type":"ContainerStarted","Data":"689afc1dac01728ffadd390c7a2587aba6197cf23819976c42d270443708afc0"} Mar 12 14:18:03 crc kubenswrapper[4778]: I0312 14:18:03.397278 4778 generic.go:334] "Generic (PLEG): container finished" podID="f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9" containerID="b15f85572ed50fa6f5f1417355d5cdd391ae57e91aab8027f8febe9070bb5ec6" exitCode=0 Mar 12 14:18:03 crc kubenswrapper[4778]: I0312 14:18:03.397394 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" event={"ID":"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9","Type":"ContainerDied","Data":"b15f85572ed50fa6f5f1417355d5cdd391ae57e91aab8027f8febe9070bb5ec6"} Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.077928 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.165874 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd9r9\" (UniqueName: \"kubernetes.io/projected/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9-kube-api-access-xd9r9\") pod \"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9\" (UID: \"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9\") " Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.172796 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9-kube-api-access-xd9r9" (OuterVolumeSpecName: "kube-api-access-xd9r9") pod "f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9" (UID: "f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9"). InnerVolumeSpecName "kube-api-access-xd9r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.268944 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd9r9\" (UniqueName: \"kubernetes.io/projected/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9-kube-api-access-xd9r9\") on node \"crc\" DevicePath \"\"" Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.417378 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" event={"ID":"f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9","Type":"ContainerDied","Data":"689afc1dac01728ffadd390c7a2587aba6197cf23819976c42d270443708afc0"} Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.417426 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="689afc1dac01728ffadd390c7a2587aba6197cf23819976c42d270443708afc0" Mar 12 14:18:05 crc kubenswrapper[4778]: I0312 14:18:05.417465 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555418-xsl6h" Mar 12 14:18:06 crc kubenswrapper[4778]: I0312 14:18:06.150460 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555412-nvskv"] Mar 12 14:18:06 crc kubenswrapper[4778]: I0312 14:18:06.160256 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555412-nvskv"] Mar 12 14:18:06 crc kubenswrapper[4778]: I0312 14:18:06.266442 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2a4f01e-04c1-43b0-8858-2d2334a828e5" path="/var/lib/kubelet/pods/f2a4f01e-04c1-43b0-8858-2d2334a828e5/volumes" Mar 12 14:18:21 crc kubenswrapper[4778]: I0312 14:18:21.273129 4778 scope.go:117] "RemoveContainer" containerID="a0fce8e55d131bb482515dc65d16783265a86aa260db0e75ebb8541d77da26bd" Mar 12 14:18:28 crc kubenswrapper[4778]: I0312 14:18:28.557678 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:18:28 crc kubenswrapper[4778]: I0312 14:18:28.558304 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.279673 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c6zkg"] Mar 12 14:18:58 crc kubenswrapper[4778]: E0312 14:18:58.280871 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9" containerName="oc" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.280888 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9" containerName="oc" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.281136 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9" containerName="oc" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.282788 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.310446 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6zkg"] Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.369175 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-catalog-content\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.369317 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-utilities\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.369435 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w46fk\" (UniqueName: \"kubernetes.io/projected/dfbf7e11-e585-48c4-a038-1a642b34bf20-kube-api-access-w46fk\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.470965 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w46fk\" (UniqueName: \"kubernetes.io/projected/dfbf7e11-e585-48c4-a038-1a642b34bf20-kube-api-access-w46fk\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.471038 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-catalog-content\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.471118 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-utilities\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.471644 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-catalog-content\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.471700 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-utilities\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.496553 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w46fk\" (UniqueName: \"kubernetes.io/projected/dfbf7e11-e585-48c4-a038-1a642b34bf20-kube-api-access-w46fk\") pod \"redhat-marketplace-c6zkg\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.557492 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.557810 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:18:58 crc kubenswrapper[4778]: I0312 14:18:58.606811 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:18:59 crc kubenswrapper[4778]: I0312 14:18:59.073059 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6zkg"] Mar 12 14:18:59 crc kubenswrapper[4778]: I0312 14:18:59.881648 4778 generic.go:334] "Generic (PLEG): container finished" podID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerID="e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85" exitCode=0 Mar 12 14:18:59 crc kubenswrapper[4778]: I0312 14:18:59.881821 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6zkg" event={"ID":"dfbf7e11-e585-48c4-a038-1a642b34bf20","Type":"ContainerDied","Data":"e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85"} Mar 12 14:18:59 crc kubenswrapper[4778]: I0312 14:18:59.881931 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6zkg" event={"ID":"dfbf7e11-e585-48c4-a038-1a642b34bf20","Type":"ContainerStarted","Data":"5f30e337fb736bed955aa279f581a9d18d0fa917fb0681de958f108990939dea"} Mar 12 14:19:01 crc kubenswrapper[4778]: E0312 14:19:01.547175 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfbf7e11_e585_48c4_a038_1a642b34bf20.slice/crio-conmon-b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfbf7e11_e585_48c4_a038_1a642b34bf20.slice/crio-b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1.scope\": RecentStats: unable to find data in memory cache]" Mar 12 14:19:01 crc kubenswrapper[4778]: I0312 14:19:01.902317 4778 generic.go:334] "Generic (PLEG): container finished" podID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerID="b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1" exitCode=0 Mar 12 14:19:01 crc kubenswrapper[4778]: I0312 14:19:01.902392 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6zkg" event={"ID":"dfbf7e11-e585-48c4-a038-1a642b34bf20","Type":"ContainerDied","Data":"b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1"} Mar 12 14:19:01 crc kubenswrapper[4778]: I0312 14:19:01.904831 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:19:02 crc kubenswrapper[4778]: I0312 14:19:02.914813 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6zkg" event={"ID":"dfbf7e11-e585-48c4-a038-1a642b34bf20","Type":"ContainerStarted","Data":"27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9"} Mar 12 14:19:08 crc kubenswrapper[4778]: I0312 14:19:08.607735 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:19:08 crc kubenswrapper[4778]: I0312 14:19:08.608030 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:19:08 crc kubenswrapper[4778]: I0312 14:19:08.762576 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:19:08 crc kubenswrapper[4778]: I0312 14:19:08.788368 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c6zkg" podStartSLOduration=8.282948396 podStartE2EDuration="10.788351047s" podCreationTimestamp="2026-03-12 14:18:58 +0000 UTC" firstStartedPulling="2026-03-12 14:18:59.884095161 +0000 UTC m=+4158.332790557" lastFinishedPulling="2026-03-12 14:19:02.389497812 +0000 UTC m=+4160.838193208" observedRunningTime="2026-03-12 14:19:02.946936592 +0000 UTC m=+4161.395631988" watchObservedRunningTime="2026-03-12 14:19:08.788351047 +0000 UTC m=+4167.237046443" Mar 12 14:19:09 crc kubenswrapper[4778]: I0312 14:19:09.020794 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:19:09 crc kubenswrapper[4778]: I0312 14:19:09.085549 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6zkg"] Mar 12 14:19:10 crc kubenswrapper[4778]: I0312 14:19:10.987137 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c6zkg" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="registry-server" containerID="cri-o://27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9" gracePeriod=2 Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.634136 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.763789 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w46fk\" (UniqueName: \"kubernetes.io/projected/dfbf7e11-e585-48c4-a038-1a642b34bf20-kube-api-access-w46fk\") pod \"dfbf7e11-e585-48c4-a038-1a642b34bf20\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.764292 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-catalog-content\") pod \"dfbf7e11-e585-48c4-a038-1a642b34bf20\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.764417 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-utilities\") pod \"dfbf7e11-e585-48c4-a038-1a642b34bf20\" (UID: \"dfbf7e11-e585-48c4-a038-1a642b34bf20\") " Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.765934 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-utilities" (OuterVolumeSpecName: "utilities") pod "dfbf7e11-e585-48c4-a038-1a642b34bf20" (UID: "dfbf7e11-e585-48c4-a038-1a642b34bf20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.777273 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfbf7e11-e585-48c4-a038-1a642b34bf20-kube-api-access-w46fk" (OuterVolumeSpecName: "kube-api-access-w46fk") pod "dfbf7e11-e585-48c4-a038-1a642b34bf20" (UID: "dfbf7e11-e585-48c4-a038-1a642b34bf20"). InnerVolumeSpecName "kube-api-access-w46fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.867547 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w46fk\" (UniqueName: \"kubernetes.io/projected/dfbf7e11-e585-48c4-a038-1a642b34bf20-kube-api-access-w46fk\") on node \"crc\" DevicePath \"\"" Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.867818 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.899166 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfbf7e11-e585-48c4-a038-1a642b34bf20" (UID: "dfbf7e11-e585-48c4-a038-1a642b34bf20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:19:11 crc kubenswrapper[4778]: I0312 14:19:11.969492 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfbf7e11-e585-48c4-a038-1a642b34bf20-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.000907 4778 generic.go:334] "Generic (PLEG): container finished" podID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerID="27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9" exitCode=0 Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.000985 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6zkg" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.001012 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6zkg" event={"ID":"dfbf7e11-e585-48c4-a038-1a642b34bf20","Type":"ContainerDied","Data":"27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9"} Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.002205 4778 scope.go:117] "RemoveContainer" containerID="27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.002133 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6zkg" event={"ID":"dfbf7e11-e585-48c4-a038-1a642b34bf20","Type":"ContainerDied","Data":"5f30e337fb736bed955aa279f581a9d18d0fa917fb0681de958f108990939dea"} Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.024150 4778 scope.go:117] "RemoveContainer" containerID="b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.046739 4778 scope.go:117] "RemoveContainer" containerID="e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.052995 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6zkg"] Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.062876 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6zkg"] Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.098036 4778 scope.go:117] "RemoveContainer" containerID="27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9" Mar 12 14:19:12 crc kubenswrapper[4778]: E0312 14:19:12.098626 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9\": container with ID starting with 27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9 not found: ID does not exist" containerID="27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.098665 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9"} err="failed to get container status \"27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9\": rpc error: code = NotFound desc = could not find container \"27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9\": container with ID starting with 27a67aa78ed15bca7e6416fb7234b4c96dc7d3121f30ecdc6273108b57e425c9 not found: ID does not exist" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.098687 4778 scope.go:117] "RemoveContainer" containerID="b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1" Mar 12 14:19:12 crc kubenswrapper[4778]: E0312 14:19:12.099655 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1\": container with ID starting with b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1 not found: ID does not exist" containerID="b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.099865 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1"} err="failed to get container status \"b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1\": rpc error: code = NotFound desc = could not find container \"b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1\": container with ID starting with b857541cbcebd77863643bfc0e6b81bac07a2dd8aa0275412e4a79c4af5595d1 not found: ID does not exist" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.099974 4778 scope.go:117] "RemoveContainer" containerID="e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85" Mar 12 14:19:12 crc kubenswrapper[4778]: E0312 14:19:12.100773 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85\": container with ID starting with e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85 not found: ID does not exist" containerID="e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.100816 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85"} err="failed to get container status \"e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85\": rpc error: code = NotFound desc = could not find container \"e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85\": container with ID starting with e6a04774d5c089a6caac52b6fd1cd02fef3c72791f4c34122389109951297e85 not found: ID does not exist" Mar 12 14:19:12 crc kubenswrapper[4778]: I0312 14:19:12.264707 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" path="/var/lib/kubelet/pods/dfbf7e11-e585-48c4-a038-1a642b34bf20/volumes" Mar 12 14:19:28 crc kubenswrapper[4778]: I0312 14:19:28.557513 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:19:28 crc kubenswrapper[4778]: I0312 14:19:28.558116 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:19:28 crc kubenswrapper[4778]: I0312 14:19:28.558197 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:19:28 crc kubenswrapper[4778]: I0312 14:19:28.559018 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a6494f4559bd62f54e1e656b9c39bd8218ddcffc6f2d4766fd788af23c632a2c"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:19:28 crc kubenswrapper[4778]: I0312 14:19:28.559078 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://a6494f4559bd62f54e1e656b9c39bd8218ddcffc6f2d4766fd788af23c632a2c" gracePeriod=600 Mar 12 14:19:29 crc kubenswrapper[4778]: I0312 14:19:29.159649 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="a6494f4559bd62f54e1e656b9c39bd8218ddcffc6f2d4766fd788af23c632a2c" exitCode=0 Mar 12 14:19:29 crc kubenswrapper[4778]: I0312 14:19:29.159957 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"a6494f4559bd62f54e1e656b9c39bd8218ddcffc6f2d4766fd788af23c632a2c"} Mar 12 14:19:29 crc kubenswrapper[4778]: I0312 14:19:29.159986 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a"} Mar 12 14:19:29 crc kubenswrapper[4778]: I0312 14:19:29.160001 4778 scope.go:117] "RemoveContainer" containerID="1f141018aeb4c8c1d3d00926888126781b45815cde38ed496c177b71b2ba7fd2" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.152003 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555420-vqx98"] Mar 12 14:20:00 crc kubenswrapper[4778]: E0312 14:20:00.153216 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="extract-utilities" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.153235 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="extract-utilities" Mar 12 14:20:00 crc kubenswrapper[4778]: E0312 14:20:00.153260 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="extract-content" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.153268 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="extract-content" Mar 12 14:20:00 crc kubenswrapper[4778]: E0312 14:20:00.153279 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="registry-server" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.153287 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="registry-server" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.153548 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfbf7e11-e585-48c4-a038-1a642b34bf20" containerName="registry-server" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.154434 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.160682 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.161261 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.161402 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.161451 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555420-vqx98"] Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.209999 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hd6z\" (UniqueName: \"kubernetes.io/projected/3d3b2fac-f000-4f5a-b253-e54ae85d507f-kube-api-access-6hd6z\") pod \"auto-csr-approver-29555420-vqx98\" (UID: \"3d3b2fac-f000-4f5a-b253-e54ae85d507f\") " pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.312008 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hd6z\" (UniqueName: \"kubernetes.io/projected/3d3b2fac-f000-4f5a-b253-e54ae85d507f-kube-api-access-6hd6z\") pod \"auto-csr-approver-29555420-vqx98\" (UID: \"3d3b2fac-f000-4f5a-b253-e54ae85d507f\") " pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.332012 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hd6z\" (UniqueName: \"kubernetes.io/projected/3d3b2fac-f000-4f5a-b253-e54ae85d507f-kube-api-access-6hd6z\") pod \"auto-csr-approver-29555420-vqx98\" (UID: \"3d3b2fac-f000-4f5a-b253-e54ae85d507f\") " pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.475535 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:00 crc kubenswrapper[4778]: I0312 14:20:00.906898 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555420-vqx98"] Mar 12 14:20:01 crc kubenswrapper[4778]: I0312 14:20:01.460571 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555420-vqx98" event={"ID":"3d3b2fac-f000-4f5a-b253-e54ae85d507f","Type":"ContainerStarted","Data":"5ebb0322b4e511455b511c13308ff5a03d744e0670cff67ace653f8b35850e2c"} Mar 12 14:20:03 crc kubenswrapper[4778]: I0312 14:20:03.480476 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555420-vqx98" event={"ID":"3d3b2fac-f000-4f5a-b253-e54ae85d507f","Type":"ContainerStarted","Data":"45d8e22a4c6b9a2b198c09597b6bc6f24b127ce1a5abca778cee677c28671528"} Mar 12 14:20:03 crc kubenswrapper[4778]: I0312 14:20:03.504618 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555420-vqx98" podStartSLOduration=2.254114855 podStartE2EDuration="3.504594157s" podCreationTimestamp="2026-03-12 14:20:00 +0000 UTC" firstStartedPulling="2026-03-12 14:20:01.234960116 +0000 UTC m=+4219.683655522" lastFinishedPulling="2026-03-12 14:20:02.485439428 +0000 UTC m=+4220.934134824" observedRunningTime="2026-03-12 14:20:03.49695438 +0000 UTC m=+4221.945649776" watchObservedRunningTime="2026-03-12 14:20:03.504594157 +0000 UTC m=+4221.953289553" Mar 12 14:20:04 crc kubenswrapper[4778]: I0312 14:20:04.492517 4778 generic.go:334] "Generic (PLEG): container finished" podID="3d3b2fac-f000-4f5a-b253-e54ae85d507f" containerID="45d8e22a4c6b9a2b198c09597b6bc6f24b127ce1a5abca778cee677c28671528" exitCode=0 Mar 12 14:20:04 crc kubenswrapper[4778]: I0312 14:20:04.492619 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555420-vqx98" event={"ID":"3d3b2fac-f000-4f5a-b253-e54ae85d507f","Type":"ContainerDied","Data":"45d8e22a4c6b9a2b198c09597b6bc6f24b127ce1a5abca778cee677c28671528"} Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.102927 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.133163 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hd6z\" (UniqueName: \"kubernetes.io/projected/3d3b2fac-f000-4f5a-b253-e54ae85d507f-kube-api-access-6hd6z\") pod \"3d3b2fac-f000-4f5a-b253-e54ae85d507f\" (UID: \"3d3b2fac-f000-4f5a-b253-e54ae85d507f\") " Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.148370 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d3b2fac-f000-4f5a-b253-e54ae85d507f-kube-api-access-6hd6z" (OuterVolumeSpecName: "kube-api-access-6hd6z") pod "3d3b2fac-f000-4f5a-b253-e54ae85d507f" (UID: "3d3b2fac-f000-4f5a-b253-e54ae85d507f"). InnerVolumeSpecName "kube-api-access-6hd6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.235430 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hd6z\" (UniqueName: \"kubernetes.io/projected/3d3b2fac-f000-4f5a-b253-e54ae85d507f-kube-api-access-6hd6z\") on node \"crc\" DevicePath \"\"" Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.514291 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555420-vqx98" event={"ID":"3d3b2fac-f000-4f5a-b253-e54ae85d507f","Type":"ContainerDied","Data":"5ebb0322b4e511455b511c13308ff5a03d744e0670cff67ace653f8b35850e2c"} Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.514333 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ebb0322b4e511455b511c13308ff5a03d744e0670cff67ace653f8b35850e2c" Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.514410 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555420-vqx98" Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.585985 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555414-sk27k"] Mar 12 14:20:06 crc kubenswrapper[4778]: I0312 14:20:06.595621 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555414-sk27k"] Mar 12 14:20:08 crc kubenswrapper[4778]: I0312 14:20:08.265618 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="066c6c88-5aea-4678-88a0-ec5c556ee008" path="/var/lib/kubelet/pods/066c6c88-5aea-4678-88a0-ec5c556ee008/volumes" Mar 12 14:20:21 crc kubenswrapper[4778]: I0312 14:20:21.377532 4778 scope.go:117] "RemoveContainer" containerID="3925769f3c54add574a18597a06eea490ae5d1cab077561f5ae8b471c0db5519" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.786058 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wgfdw"] Mar 12 14:21:15 crc kubenswrapper[4778]: E0312 14:21:15.786955 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3b2fac-f000-4f5a-b253-e54ae85d507f" containerName="oc" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.786968 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3b2fac-f000-4f5a-b253-e54ae85d507f" containerName="oc" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.787292 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d3b2fac-f000-4f5a-b253-e54ae85d507f" containerName="oc" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.788672 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.800414 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wgfdw"] Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.869140 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-catalog-content\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.869221 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlsbx\" (UniqueName: \"kubernetes.io/projected/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-kube-api-access-rlsbx\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.869462 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-utilities\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.971685 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-catalog-content\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.971747 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlsbx\" (UniqueName: \"kubernetes.io/projected/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-kube-api-access-rlsbx\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.971903 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-utilities\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.972348 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-catalog-content\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.972380 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-utilities\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:15 crc kubenswrapper[4778]: I0312 14:21:15.999703 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlsbx\" (UniqueName: \"kubernetes.io/projected/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-kube-api-access-rlsbx\") pod \"redhat-operators-wgfdw\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:16 crc kubenswrapper[4778]: I0312 14:21:16.109476 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:16 crc kubenswrapper[4778]: I0312 14:21:16.663339 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wgfdw"] Mar 12 14:21:17 crc kubenswrapper[4778]: I0312 14:21:17.125925 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerID="e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93" exitCode=0 Mar 12 14:21:17 crc kubenswrapper[4778]: I0312 14:21:17.125983 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerDied","Data":"e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93"} Mar 12 14:21:17 crc kubenswrapper[4778]: I0312 14:21:17.126229 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerStarted","Data":"089a716699a43fe932443bd8a22b6bddce9e43f3949e83593a8c791039aa04a6"} Mar 12 14:21:18 crc kubenswrapper[4778]: I0312 14:21:18.138575 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerStarted","Data":"58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d"} Mar 12 14:21:23 crc kubenswrapper[4778]: I0312 14:21:23.183883 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerID="58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d" exitCode=0 Mar 12 14:21:23 crc kubenswrapper[4778]: I0312 14:21:23.184084 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerDied","Data":"58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d"} Mar 12 14:21:24 crc kubenswrapper[4778]: I0312 14:21:24.194438 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerStarted","Data":"db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3"} Mar 12 14:21:24 crc kubenswrapper[4778]: I0312 14:21:24.222558 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wgfdw" podStartSLOduration=2.516719762 podStartE2EDuration="9.222536099s" podCreationTimestamp="2026-03-12 14:21:15 +0000 UTC" firstStartedPulling="2026-03-12 14:21:17.127568465 +0000 UTC m=+4295.576263861" lastFinishedPulling="2026-03-12 14:21:23.833384802 +0000 UTC m=+4302.282080198" observedRunningTime="2026-03-12 14:21:24.215679284 +0000 UTC m=+4302.664374700" watchObservedRunningTime="2026-03-12 14:21:24.222536099 +0000 UTC m=+4302.671231495" Mar 12 14:21:26 crc kubenswrapper[4778]: I0312 14:21:26.110343 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:26 crc kubenswrapper[4778]: I0312 14:21:26.110694 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:27 crc kubenswrapper[4778]: I0312 14:21:27.160804 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wgfdw" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" probeResult="failure" output=< Mar 12 14:21:27 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:21:27 crc kubenswrapper[4778]: > Mar 12 14:21:28 crc kubenswrapper[4778]: I0312 14:21:28.557373 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:21:28 crc kubenswrapper[4778]: I0312 14:21:28.557739 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:21:37 crc kubenswrapper[4778]: I0312 14:21:37.157265 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wgfdw" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" probeResult="failure" output=< Mar 12 14:21:37 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:21:37 crc kubenswrapper[4778]: > Mar 12 14:21:47 crc kubenswrapper[4778]: I0312 14:21:47.166474 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wgfdw" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" probeResult="failure" output=< Mar 12 14:21:47 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:21:47 crc kubenswrapper[4778]: > Mar 12 14:21:56 crc kubenswrapper[4778]: I0312 14:21:56.164256 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:56 crc kubenswrapper[4778]: I0312 14:21:56.229599 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:56 crc kubenswrapper[4778]: I0312 14:21:56.402016 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wgfdw"] Mar 12 14:21:57 crc kubenswrapper[4778]: I0312 14:21:57.503273 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wgfdw" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" containerID="cri-o://db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3" gracePeriod=2 Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.296688 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.343872 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-catalog-content\") pod \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.343950 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlsbx\" (UniqueName: \"kubernetes.io/projected/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-kube-api-access-rlsbx\") pod \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.343975 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-utilities\") pod \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\" (UID: \"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe\") " Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.346511 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-utilities" (OuterVolumeSpecName: "utilities") pod "0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" (UID: "0bc2634e-eaa3-4b87-977d-000a3fe0ccbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.352424 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-kube-api-access-rlsbx" (OuterVolumeSpecName: "kube-api-access-rlsbx") pod "0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" (UID: "0bc2634e-eaa3-4b87-977d-000a3fe0ccbe"). InnerVolumeSpecName "kube-api-access-rlsbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.447275 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlsbx\" (UniqueName: \"kubernetes.io/projected/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-kube-api-access-rlsbx\") on node \"crc\" DevicePath \"\"" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.447318 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.502072 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" (UID: "0bc2634e-eaa3-4b87-977d-000a3fe0ccbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.515715 4778 generic.go:334] "Generic (PLEG): container finished" podID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerID="db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3" exitCode=0 Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.515769 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerDied","Data":"db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3"} Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.515801 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgfdw" event={"ID":"0bc2634e-eaa3-4b87-977d-000a3fe0ccbe","Type":"ContainerDied","Data":"089a716699a43fe932443bd8a22b6bddce9e43f3949e83593a8c791039aa04a6"} Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.515819 4778 scope.go:117] "RemoveContainer" containerID="db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.515856 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgfdw" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.537836 4778 scope.go:117] "RemoveContainer" containerID="58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.550344 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.551596 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wgfdw"] Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.557880 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.557965 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.560809 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wgfdw"] Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.576730 4778 scope.go:117] "RemoveContainer" containerID="e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.615398 4778 scope.go:117] "RemoveContainer" containerID="db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3" Mar 12 14:21:58 crc kubenswrapper[4778]: E0312 14:21:58.616053 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3\": container with ID starting with db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3 not found: ID does not exist" containerID="db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.616087 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3"} err="failed to get container status \"db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3\": rpc error: code = NotFound desc = could not find container \"db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3\": container with ID starting with db8b5f8281bbd18e975804d8003a6e06615d4d5f74acea8bc7b682044597e6e3 not found: ID does not exist" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.616108 4778 scope.go:117] "RemoveContainer" containerID="58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d" Mar 12 14:21:58 crc kubenswrapper[4778]: E0312 14:21:58.616615 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d\": container with ID starting with 58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d not found: ID does not exist" containerID="58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.616635 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d"} err="failed to get container status \"58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d\": rpc error: code = NotFound desc = could not find container \"58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d\": container with ID starting with 58bb40fbf4d275731a341f0ca74d14e00964d2cc83f015da64db3332843f4f6d not found: ID does not exist" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.616646 4778 scope.go:117] "RemoveContainer" containerID="e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93" Mar 12 14:21:58 crc kubenswrapper[4778]: E0312 14:21:58.617008 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93\": container with ID starting with e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93 not found: ID does not exist" containerID="e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93" Mar 12 14:21:58 crc kubenswrapper[4778]: I0312 14:21:58.617043 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93"} err="failed to get container status \"e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93\": rpc error: code = NotFound desc = could not find container \"e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93\": container with ID starting with e16cdbdf5f70c6a7cdf18ed234bc9607178e963e177bd90f1f9a079c2ffdca93 not found: ID does not exist" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.152727 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555422-kw6c8"] Mar 12 14:22:00 crc kubenswrapper[4778]: E0312 14:22:00.153690 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="extract-utilities" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.153710 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="extract-utilities" Mar 12 14:22:00 crc kubenswrapper[4778]: E0312 14:22:00.153742 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="extract-content" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.153750 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="extract-content" Mar 12 14:22:00 crc kubenswrapper[4778]: E0312 14:22:00.153764 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.153774 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.153991 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" containerName="registry-server" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.154835 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.160349 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.160548 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.160767 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.164021 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555422-kw6c8"] Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.266900 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc2634e-eaa3-4b87-977d-000a3fe0ccbe" path="/var/lib/kubelet/pods/0bc2634e-eaa3-4b87-977d-000a3fe0ccbe/volumes" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.302112 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7tq9\" (UniqueName: \"kubernetes.io/projected/ecab0458-0ee6-4672-bd27-4c8aae8427bb-kube-api-access-j7tq9\") pod \"auto-csr-approver-29555422-kw6c8\" (UID: \"ecab0458-0ee6-4672-bd27-4c8aae8427bb\") " pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.404395 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7tq9\" (UniqueName: \"kubernetes.io/projected/ecab0458-0ee6-4672-bd27-4c8aae8427bb-kube-api-access-j7tq9\") pod \"auto-csr-approver-29555422-kw6c8\" (UID: \"ecab0458-0ee6-4672-bd27-4c8aae8427bb\") " pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.427384 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7tq9\" (UniqueName: \"kubernetes.io/projected/ecab0458-0ee6-4672-bd27-4c8aae8427bb-kube-api-access-j7tq9\") pod \"auto-csr-approver-29555422-kw6c8\" (UID: \"ecab0458-0ee6-4672-bd27-4c8aae8427bb\") " pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.490229 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:00 crc kubenswrapper[4778]: I0312 14:22:00.963403 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555422-kw6c8"] Mar 12 14:22:01 crc kubenswrapper[4778]: I0312 14:22:01.548005 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" event={"ID":"ecab0458-0ee6-4672-bd27-4c8aae8427bb","Type":"ContainerStarted","Data":"eaa71bc1a9088796769c4f5fda53031c0ace72bdbec06914242d1c22e12b5576"} Mar 12 14:22:04 crc kubenswrapper[4778]: I0312 14:22:04.575234 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" event={"ID":"ecab0458-0ee6-4672-bd27-4c8aae8427bb","Type":"ContainerStarted","Data":"460cb8eb02f9333998d559fe47fe50a7beb133708302defa156052aac3033d0e"} Mar 12 14:22:05 crc kubenswrapper[4778]: I0312 14:22:05.586216 4778 generic.go:334] "Generic (PLEG): container finished" podID="ecab0458-0ee6-4672-bd27-4c8aae8427bb" containerID="460cb8eb02f9333998d559fe47fe50a7beb133708302defa156052aac3033d0e" exitCode=0 Mar 12 14:22:05 crc kubenswrapper[4778]: I0312 14:22:05.586335 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" event={"ID":"ecab0458-0ee6-4672-bd27-4c8aae8427bb","Type":"ContainerDied","Data":"460cb8eb02f9333998d559fe47fe50a7beb133708302defa156052aac3033d0e"} Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.113709 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.252514 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7tq9\" (UniqueName: \"kubernetes.io/projected/ecab0458-0ee6-4672-bd27-4c8aae8427bb-kube-api-access-j7tq9\") pod \"ecab0458-0ee6-4672-bd27-4c8aae8427bb\" (UID: \"ecab0458-0ee6-4672-bd27-4c8aae8427bb\") " Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.259440 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecab0458-0ee6-4672-bd27-4c8aae8427bb-kube-api-access-j7tq9" (OuterVolumeSpecName: "kube-api-access-j7tq9") pod "ecab0458-0ee6-4672-bd27-4c8aae8427bb" (UID: "ecab0458-0ee6-4672-bd27-4c8aae8427bb"). InnerVolumeSpecName "kube-api-access-j7tq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.354886 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7tq9\" (UniqueName: \"kubernetes.io/projected/ecab0458-0ee6-4672-bd27-4c8aae8427bb-kube-api-access-j7tq9\") on node \"crc\" DevicePath \"\"" Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.608505 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" event={"ID":"ecab0458-0ee6-4672-bd27-4c8aae8427bb","Type":"ContainerDied","Data":"eaa71bc1a9088796769c4f5fda53031c0ace72bdbec06914242d1c22e12b5576"} Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.608557 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaa71bc1a9088796769c4f5fda53031c0ace72bdbec06914242d1c22e12b5576" Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.608593 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555422-kw6c8" Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.658121 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555416-qx4gr"] Mar 12 14:22:07 crc kubenswrapper[4778]: I0312 14:22:07.674480 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555416-qx4gr"] Mar 12 14:22:08 crc kubenswrapper[4778]: I0312 14:22:08.263784 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43680ad6-62d2-4e00-a38b-e87d712af9a7" path="/var/lib/kubelet/pods/43680ad6-62d2-4e00-a38b-e87d712af9a7/volumes" Mar 12 14:22:21 crc kubenswrapper[4778]: I0312 14:22:21.613970 4778 scope.go:117] "RemoveContainer" containerID="95522e18d302b349263bfd01c0f317ec6a698231318f520fcd1ed51d7aa504cc" Mar 12 14:22:28 crc kubenswrapper[4778]: I0312 14:22:28.557622 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:22:28 crc kubenswrapper[4778]: I0312 14:22:28.558254 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:22:28 crc kubenswrapper[4778]: I0312 14:22:28.558312 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:22:28 crc kubenswrapper[4778]: I0312 14:22:28.559101 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:22:28 crc kubenswrapper[4778]: I0312 14:22:28.559152 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" gracePeriod=600 Mar 12 14:22:28 crc kubenswrapper[4778]: E0312 14:22:28.679997 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:22:29 crc kubenswrapper[4778]: I0312 14:22:29.534032 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" exitCode=0 Mar 12 14:22:29 crc kubenswrapper[4778]: I0312 14:22:29.534080 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a"} Mar 12 14:22:29 crc kubenswrapper[4778]: I0312 14:22:29.534118 4778 scope.go:117] "RemoveContainer" containerID="a6494f4559bd62f54e1e656b9c39bd8218ddcffc6f2d4766fd788af23c632a2c" Mar 12 14:22:29 crc kubenswrapper[4778]: I0312 14:22:29.534880 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:22:29 crc kubenswrapper[4778]: E0312 14:22:29.535269 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:22:40 crc kubenswrapper[4778]: I0312 14:22:40.253675 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:22:40 crc kubenswrapper[4778]: E0312 14:22:40.254504 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:22:53 crc kubenswrapper[4778]: I0312 14:22:53.254428 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:22:53 crc kubenswrapper[4778]: E0312 14:22:53.255362 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:23:04 crc kubenswrapper[4778]: I0312 14:23:04.253517 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:23:04 crc kubenswrapper[4778]: E0312 14:23:04.254315 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:23:15 crc kubenswrapper[4778]: I0312 14:23:15.253986 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:23:15 crc kubenswrapper[4778]: E0312 14:23:15.254815 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:23:26 crc kubenswrapper[4778]: I0312 14:23:26.253511 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:23:26 crc kubenswrapper[4778]: E0312 14:23:26.254214 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:23:37 crc kubenswrapper[4778]: I0312 14:23:37.253794 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:23:37 crc kubenswrapper[4778]: E0312 14:23:37.254497 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:23:52 crc kubenswrapper[4778]: I0312 14:23:52.260741 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:23:52 crc kubenswrapper[4778]: E0312 14:23:52.262245 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.153033 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555424-5hp4x"] Mar 12 14:24:00 crc kubenswrapper[4778]: E0312 14:24:00.154316 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecab0458-0ee6-4672-bd27-4c8aae8427bb" containerName="oc" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.154332 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecab0458-0ee6-4672-bd27-4c8aae8427bb" containerName="oc" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.154660 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecab0458-0ee6-4672-bd27-4c8aae8427bb" containerName="oc" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.155513 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.161418 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.161432 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.162617 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.167221 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555424-5hp4x"] Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.261897 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msg5g\" (UniqueName: \"kubernetes.io/projected/78c94941-b604-4fc0-b7b3-0b6749bbf233-kube-api-access-msg5g\") pod \"auto-csr-approver-29555424-5hp4x\" (UID: \"78c94941-b604-4fc0-b7b3-0b6749bbf233\") " pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.363489 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msg5g\" (UniqueName: \"kubernetes.io/projected/78c94941-b604-4fc0-b7b3-0b6749bbf233-kube-api-access-msg5g\") pod \"auto-csr-approver-29555424-5hp4x\" (UID: \"78c94941-b604-4fc0-b7b3-0b6749bbf233\") " pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.385029 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msg5g\" (UniqueName: \"kubernetes.io/projected/78c94941-b604-4fc0-b7b3-0b6749bbf233-kube-api-access-msg5g\") pod \"auto-csr-approver-29555424-5hp4x\" (UID: \"78c94941-b604-4fc0-b7b3-0b6749bbf233\") " pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.479409 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:00 crc kubenswrapper[4778]: I0312 14:24:00.995260 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555424-5hp4x"] Mar 12 14:24:00 crc kubenswrapper[4778]: W0312 14:24:00.997586 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78c94941_b604_4fc0_b7b3_0b6749bbf233.slice/crio-6b42bc5d5bf8ed9b31884ad16ead36dae0cc5b8a58690359c02f1ddd78126a2f WatchSource:0}: Error finding container 6b42bc5d5bf8ed9b31884ad16ead36dae0cc5b8a58690359c02f1ddd78126a2f: Status 404 returned error can't find the container with id 6b42bc5d5bf8ed9b31884ad16ead36dae0cc5b8a58690359c02f1ddd78126a2f Mar 12 14:24:01 crc kubenswrapper[4778]: I0312 14:24:01.046058 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" event={"ID":"78c94941-b604-4fc0-b7b3-0b6749bbf233","Type":"ContainerStarted","Data":"6b42bc5d5bf8ed9b31884ad16ead36dae0cc5b8a58690359c02f1ddd78126a2f"} Mar 12 14:24:03 crc kubenswrapper[4778]: I0312 14:24:03.068092 4778 generic.go:334] "Generic (PLEG): container finished" podID="78c94941-b604-4fc0-b7b3-0b6749bbf233" containerID="a1c2cc27e654689e4f136031bb0129f78011ff7542f974149c494096b483a2a2" exitCode=0 Mar 12 14:24:03 crc kubenswrapper[4778]: I0312 14:24:03.068240 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" event={"ID":"78c94941-b604-4fc0-b7b3-0b6749bbf233","Type":"ContainerDied","Data":"a1c2cc27e654689e4f136031bb0129f78011ff7542f974149c494096b483a2a2"} Mar 12 14:24:04 crc kubenswrapper[4778]: I0312 14:24:04.527425 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:04 crc kubenswrapper[4778]: I0312 14:24:04.653687 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msg5g\" (UniqueName: \"kubernetes.io/projected/78c94941-b604-4fc0-b7b3-0b6749bbf233-kube-api-access-msg5g\") pod \"78c94941-b604-4fc0-b7b3-0b6749bbf233\" (UID: \"78c94941-b604-4fc0-b7b3-0b6749bbf233\") " Mar 12 14:24:04 crc kubenswrapper[4778]: I0312 14:24:04.658887 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c94941-b604-4fc0-b7b3-0b6749bbf233-kube-api-access-msg5g" (OuterVolumeSpecName: "kube-api-access-msg5g") pod "78c94941-b604-4fc0-b7b3-0b6749bbf233" (UID: "78c94941-b604-4fc0-b7b3-0b6749bbf233"). InnerVolumeSpecName "kube-api-access-msg5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:24:04 crc kubenswrapper[4778]: I0312 14:24:04.756290 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msg5g\" (UniqueName: \"kubernetes.io/projected/78c94941-b604-4fc0-b7b3-0b6749bbf233-kube-api-access-msg5g\") on node \"crc\" DevicePath \"\"" Mar 12 14:24:05 crc kubenswrapper[4778]: I0312 14:24:05.087364 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" event={"ID":"78c94941-b604-4fc0-b7b3-0b6749bbf233","Type":"ContainerDied","Data":"6b42bc5d5bf8ed9b31884ad16ead36dae0cc5b8a58690359c02f1ddd78126a2f"} Mar 12 14:24:05 crc kubenswrapper[4778]: I0312 14:24:05.087419 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b42bc5d5bf8ed9b31884ad16ead36dae0cc5b8a58690359c02f1ddd78126a2f" Mar 12 14:24:05 crc kubenswrapper[4778]: I0312 14:24:05.087429 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555424-5hp4x" Mar 12 14:24:05 crc kubenswrapper[4778]: I0312 14:24:05.254229 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:24:05 crc kubenswrapper[4778]: E0312 14:24:05.254651 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:24:05 crc kubenswrapper[4778]: I0312 14:24:05.599666 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555418-xsl6h"] Mar 12 14:24:05 crc kubenswrapper[4778]: I0312 14:24:05.610430 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555418-xsl6h"] Mar 12 14:24:06 crc kubenswrapper[4778]: I0312 14:24:06.265678 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9" path="/var/lib/kubelet/pods/f3fd5c4e-83c5-4ff0-9cb3-665ac00ec9f9/volumes" Mar 12 14:24:16 crc kubenswrapper[4778]: I0312 14:24:16.254538 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:24:16 crc kubenswrapper[4778]: E0312 14:24:16.255237 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:24:21 crc kubenswrapper[4778]: I0312 14:24:21.715875 4778 scope.go:117] "RemoveContainer" containerID="b15f85572ed50fa6f5f1417355d5cdd391ae57e91aab8027f8febe9070bb5ec6" Mar 12 14:24:29 crc kubenswrapper[4778]: I0312 14:24:29.362437 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:24:29 crc kubenswrapper[4778]: E0312 14:24:29.363330 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:24:44 crc kubenswrapper[4778]: I0312 14:24:44.254479 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:24:44 crc kubenswrapper[4778]: E0312 14:24:44.256448 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:24:56 crc kubenswrapper[4778]: I0312 14:24:56.253605 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:24:56 crc kubenswrapper[4778]: E0312 14:24:56.254415 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:25:11 crc kubenswrapper[4778]: I0312 14:25:11.253949 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:25:11 crc kubenswrapper[4778]: E0312 14:25:11.254839 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:25:22 crc kubenswrapper[4778]: I0312 14:25:22.261595 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:25:22 crc kubenswrapper[4778]: E0312 14:25:22.262434 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:25:36 crc kubenswrapper[4778]: I0312 14:25:36.255604 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:25:36 crc kubenswrapper[4778]: E0312 14:25:36.257709 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:25:51 crc kubenswrapper[4778]: I0312 14:25:51.254072 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:25:51 crc kubenswrapper[4778]: E0312 14:25:51.254862 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.143510 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2csrm"] Mar 12 14:25:55 crc kubenswrapper[4778]: E0312 14:25:55.144552 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c94941-b604-4fc0-b7b3-0b6749bbf233" containerName="oc" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.144568 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c94941-b604-4fc0-b7b3-0b6749bbf233" containerName="oc" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.144752 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c94941-b604-4fc0-b7b3-0b6749bbf233" containerName="oc" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.146278 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.179408 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc8ds\" (UniqueName: \"kubernetes.io/projected/21759d07-307f-4331-94fd-5e4720ef2b7f-kube-api-access-xc8ds\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.179529 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-utilities\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.179580 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-catalog-content\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.193260 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2csrm"] Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.281658 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-utilities\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.282063 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-catalog-content\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.282453 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc8ds\" (UniqueName: \"kubernetes.io/projected/21759d07-307f-4331-94fd-5e4720ef2b7f-kube-api-access-xc8ds\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.283398 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-utilities\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.283796 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-catalog-content\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.301143 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc8ds\" (UniqueName: \"kubernetes.io/projected/21759d07-307f-4331-94fd-5e4720ef2b7f-kube-api-access-xc8ds\") pod \"certified-operators-2csrm\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.473045 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.748131 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d5k5f"] Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.750449 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.765569 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5k5f"] Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.793149 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-catalog-content\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.793305 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-utilities\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.793434 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2x6\" (UniqueName: \"kubernetes.io/projected/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-kube-api-access-kf2x6\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.895597 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2x6\" (UniqueName: \"kubernetes.io/projected/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-kube-api-access-kf2x6\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.895660 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-catalog-content\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.895778 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-utilities\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.896299 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-utilities\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.897094 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-catalog-content\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.937594 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2x6\" (UniqueName: \"kubernetes.io/projected/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-kube-api-access-kf2x6\") pod \"community-operators-d5k5f\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:55 crc kubenswrapper[4778]: I0312 14:25:55.982677 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2csrm"] Mar 12 14:25:56 crc kubenswrapper[4778]: I0312 14:25:56.088765 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:25:56 crc kubenswrapper[4778]: I0312 14:25:56.151734 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerStarted","Data":"efd699a3abbb70926c35b02e780e646800bf810115efc6de8fa07ae4825b8be7"} Mar 12 14:25:56 crc kubenswrapper[4778]: I0312 14:25:56.682936 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5k5f"] Mar 12 14:25:57 crc kubenswrapper[4778]: I0312 14:25:57.162494 4778 generic.go:334] "Generic (PLEG): container finished" podID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerID="72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380" exitCode=0 Mar 12 14:25:57 crc kubenswrapper[4778]: I0312 14:25:57.162689 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerDied","Data":"72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380"} Mar 12 14:25:57 crc kubenswrapper[4778]: I0312 14:25:57.164525 4778 generic.go:334] "Generic (PLEG): container finished" podID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerID="9049643c60cdd1df5b7a09b9be7298790ce9d88b86df75d5b1bc6953a508f058" exitCode=0 Mar 12 14:25:57 crc kubenswrapper[4778]: I0312 14:25:57.164585 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerDied","Data":"9049643c60cdd1df5b7a09b9be7298790ce9d88b86df75d5b1bc6953a508f058"} Mar 12 14:25:57 crc kubenswrapper[4778]: I0312 14:25:57.164612 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerStarted","Data":"32a087503db88c25852e194ce746844351fc2c42e470484a64156247f43bc6ee"} Mar 12 14:25:57 crc kubenswrapper[4778]: I0312 14:25:57.164818 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:25:58 crc kubenswrapper[4778]: I0312 14:25:58.175909 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerStarted","Data":"46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b"} Mar 12 14:25:58 crc kubenswrapper[4778]: I0312 14:25:58.178730 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerStarted","Data":"089be0f6a7588cf95d5a557ff5fcbb10a29194745bfdb53617ac1aedce8267e2"} Mar 12 14:25:59 crc kubenswrapper[4778]: I0312 14:25:59.194977 4778 generic.go:334] "Generic (PLEG): container finished" podID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerID="46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b" exitCode=0 Mar 12 14:25:59 crc kubenswrapper[4778]: I0312 14:25:59.195036 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerDied","Data":"46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b"} Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.143548 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555426-btwds"] Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.145591 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.148717 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.150715 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.151435 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.166046 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555426-btwds"] Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.224173 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerStarted","Data":"e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e"} Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.228430 4778 generic.go:334] "Generic (PLEG): container finished" podID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerID="089be0f6a7588cf95d5a557ff5fcbb10a29194745bfdb53617ac1aedce8267e2" exitCode=0 Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.228477 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerDied","Data":"089be0f6a7588cf95d5a557ff5fcbb10a29194745bfdb53617ac1aedce8267e2"} Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.245126 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2csrm" podStartSLOduration=2.842717645 podStartE2EDuration="5.24510405s" podCreationTimestamp="2026-03-12 14:25:55 +0000 UTC" firstStartedPulling="2026-03-12 14:25:57.164618117 +0000 UTC m=+4575.613313513" lastFinishedPulling="2026-03-12 14:25:59.567004522 +0000 UTC m=+4578.015699918" observedRunningTime="2026-03-12 14:26:00.242430384 +0000 UTC m=+4578.691125800" watchObservedRunningTime="2026-03-12 14:26:00.24510405 +0000 UTC m=+4578.693799446" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.289372 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dchwt\" (UniqueName: \"kubernetes.io/projected/fa45db55-92a4-4a16-9455-ee110dc34fa6-kube-api-access-dchwt\") pod \"auto-csr-approver-29555426-btwds\" (UID: \"fa45db55-92a4-4a16-9455-ee110dc34fa6\") " pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.391173 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dchwt\" (UniqueName: \"kubernetes.io/projected/fa45db55-92a4-4a16-9455-ee110dc34fa6-kube-api-access-dchwt\") pod \"auto-csr-approver-29555426-btwds\" (UID: \"fa45db55-92a4-4a16-9455-ee110dc34fa6\") " pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.415592 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dchwt\" (UniqueName: \"kubernetes.io/projected/fa45db55-92a4-4a16-9455-ee110dc34fa6-kube-api-access-dchwt\") pod \"auto-csr-approver-29555426-btwds\" (UID: \"fa45db55-92a4-4a16-9455-ee110dc34fa6\") " pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.473176 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:00 crc kubenswrapper[4778]: I0312 14:26:00.952281 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555426-btwds"] Mar 12 14:26:00 crc kubenswrapper[4778]: W0312 14:26:00.955378 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa45db55_92a4_4a16_9455_ee110dc34fa6.slice/crio-1c6c5c601fa7845c122e44d0bd6f8a7b743d280b9cd09f719273c5acdf0b1363 WatchSource:0}: Error finding container 1c6c5c601fa7845c122e44d0bd6f8a7b743d280b9cd09f719273c5acdf0b1363: Status 404 returned error can't find the container with id 1c6c5c601fa7845c122e44d0bd6f8a7b743d280b9cd09f719273c5acdf0b1363 Mar 12 14:26:01 crc kubenswrapper[4778]: I0312 14:26:01.239331 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555426-btwds" event={"ID":"fa45db55-92a4-4a16-9455-ee110dc34fa6","Type":"ContainerStarted","Data":"1c6c5c601fa7845c122e44d0bd6f8a7b743d280b9cd09f719273c5acdf0b1363"} Mar 12 14:26:02 crc kubenswrapper[4778]: I0312 14:26:02.267895 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerStarted","Data":"7996f5c983dd5f46d694007de45d6212a50151bb7f287388b92acc9fa36b446c"} Mar 12 14:26:02 crc kubenswrapper[4778]: I0312 14:26:02.311478 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d5k5f" podStartSLOduration=2.771586065 podStartE2EDuration="7.311462886s" podCreationTimestamp="2026-03-12 14:25:55 +0000 UTC" firstStartedPulling="2026-03-12 14:25:57.165661277 +0000 UTC m=+4575.614356673" lastFinishedPulling="2026-03-12 14:26:01.705538078 +0000 UTC m=+4580.154233494" observedRunningTime="2026-03-12 14:26:02.30349987 +0000 UTC m=+4580.752195266" watchObservedRunningTime="2026-03-12 14:26:02.311462886 +0000 UTC m=+4580.760158282" Mar 12 14:26:03 crc kubenswrapper[4778]: I0312 14:26:03.266355 4778 generic.go:334] "Generic (PLEG): container finished" podID="fa45db55-92a4-4a16-9455-ee110dc34fa6" containerID="188dd1cb886e6788ffd8398573fda57dc92b1fe481e6f2ffdc97a0e049e9348c" exitCode=0 Mar 12 14:26:03 crc kubenswrapper[4778]: I0312 14:26:03.266462 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555426-btwds" event={"ID":"fa45db55-92a4-4a16-9455-ee110dc34fa6","Type":"ContainerDied","Data":"188dd1cb886e6788ffd8398573fda57dc92b1fe481e6f2ffdc97a0e049e9348c"} Mar 12 14:26:04 crc kubenswrapper[4778]: I0312 14:26:04.834819 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:04 crc kubenswrapper[4778]: I0312 14:26:04.896005 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dchwt\" (UniqueName: \"kubernetes.io/projected/fa45db55-92a4-4a16-9455-ee110dc34fa6-kube-api-access-dchwt\") pod \"fa45db55-92a4-4a16-9455-ee110dc34fa6\" (UID: \"fa45db55-92a4-4a16-9455-ee110dc34fa6\") " Mar 12 14:26:04 crc kubenswrapper[4778]: I0312 14:26:04.901730 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa45db55-92a4-4a16-9455-ee110dc34fa6-kube-api-access-dchwt" (OuterVolumeSpecName: "kube-api-access-dchwt") pod "fa45db55-92a4-4a16-9455-ee110dc34fa6" (UID: "fa45db55-92a4-4a16-9455-ee110dc34fa6"). InnerVolumeSpecName "kube-api-access-dchwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:26:04 crc kubenswrapper[4778]: I0312 14:26:04.998508 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dchwt\" (UniqueName: \"kubernetes.io/projected/fa45db55-92a4-4a16-9455-ee110dc34fa6-kube-api-access-dchwt\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.284046 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555426-btwds" event={"ID":"fa45db55-92a4-4a16-9455-ee110dc34fa6","Type":"ContainerDied","Data":"1c6c5c601fa7845c122e44d0bd6f8a7b743d280b9cd09f719273c5acdf0b1363"} Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.284086 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c6c5c601fa7845c122e44d0bd6f8a7b743d280b9cd09f719273c5acdf0b1363" Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.284107 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555426-btwds" Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.473788 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.473859 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.521585 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.916961 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555420-vqx98"] Mar 12 14:26:05 crc kubenswrapper[4778]: I0312 14:26:05.930270 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555420-vqx98"] Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.090923 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.090973 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.137924 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.254496 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:26:06 crc kubenswrapper[4778]: E0312 14:26:06.254966 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.267712 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d3b2fac-f000-4f5a-b253-e54ae85d507f" path="/var/lib/kubelet/pods/3d3b2fac-f000-4f5a-b253-e54ae85d507f/volumes" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.335748 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.347799 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:26:06 crc kubenswrapper[4778]: I0312 14:26:06.925417 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5k5f"] Mar 12 14:26:08 crc kubenswrapper[4778]: I0312 14:26:08.306597 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d5k5f" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="registry-server" containerID="cri-o://7996f5c983dd5f46d694007de45d6212a50151bb7f287388b92acc9fa36b446c" gracePeriod=2 Mar 12 14:26:08 crc kubenswrapper[4778]: I0312 14:26:08.732517 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2csrm"] Mar 12 14:26:08 crc kubenswrapper[4778]: I0312 14:26:08.732892 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2csrm" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="registry-server" containerID="cri-o://e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e" gracePeriod=2 Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.831537 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.906144 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc8ds\" (UniqueName: \"kubernetes.io/projected/21759d07-307f-4331-94fd-5e4720ef2b7f-kube-api-access-xc8ds\") pod \"21759d07-307f-4331-94fd-5e4720ef2b7f\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.906286 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-catalog-content\") pod \"21759d07-307f-4331-94fd-5e4720ef2b7f\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.906323 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-utilities\") pod \"21759d07-307f-4331-94fd-5e4720ef2b7f\" (UID: \"21759d07-307f-4331-94fd-5e4720ef2b7f\") " Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.907593 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-utilities" (OuterVolumeSpecName: "utilities") pod "21759d07-307f-4331-94fd-5e4720ef2b7f" (UID: "21759d07-307f-4331-94fd-5e4720ef2b7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.912406 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21759d07-307f-4331-94fd-5e4720ef2b7f-kube-api-access-xc8ds" (OuterVolumeSpecName: "kube-api-access-xc8ds") pod "21759d07-307f-4331-94fd-5e4720ef2b7f" (UID: "21759d07-307f-4331-94fd-5e4720ef2b7f"). InnerVolumeSpecName "kube-api-access-xc8ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:26:09 crc kubenswrapper[4778]: I0312 14:26:09.978628 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21759d07-307f-4331-94fd-5e4720ef2b7f" (UID: "21759d07-307f-4331-94fd-5e4720ef2b7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.008541 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc8ds\" (UniqueName: \"kubernetes.io/projected/21759d07-307f-4331-94fd-5e4720ef2b7f-kube-api-access-xc8ds\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.008579 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.008591 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21759d07-307f-4331-94fd-5e4720ef2b7f-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.324368 4778 generic.go:334] "Generic (PLEG): container finished" podID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerID="e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e" exitCode=0 Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.324667 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2csrm" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.324568 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerDied","Data":"e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e"} Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.324747 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2csrm" event={"ID":"21759d07-307f-4331-94fd-5e4720ef2b7f","Type":"ContainerDied","Data":"efd699a3abbb70926c35b02e780e646800bf810115efc6de8fa07ae4825b8be7"} Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.324796 4778 scope.go:117] "RemoveContainer" containerID="e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.330060 4778 generic.go:334] "Generic (PLEG): container finished" podID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerID="7996f5c983dd5f46d694007de45d6212a50151bb7f287388b92acc9fa36b446c" exitCode=0 Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.330112 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerDied","Data":"7996f5c983dd5f46d694007de45d6212a50151bb7f287388b92acc9fa36b446c"} Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.543270 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.557335 4778 scope.go:117] "RemoveContainer" containerID="46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.563791 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2csrm"] Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.571059 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2csrm"] Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.591084 4778 scope.go:117] "RemoveContainer" containerID="72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.620410 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf2x6\" (UniqueName: \"kubernetes.io/projected/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-kube-api-access-kf2x6\") pod \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.620959 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-catalog-content\") pod \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.621030 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-utilities\") pod \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\" (UID: \"afc8bbf3-0297-4e5a-ba30-a8fac38c3832\") " Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.621795 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-utilities" (OuterVolumeSpecName: "utilities") pod "afc8bbf3-0297-4e5a-ba30-a8fac38c3832" (UID: "afc8bbf3-0297-4e5a-ba30-a8fac38c3832"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.626106 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-kube-api-access-kf2x6" (OuterVolumeSpecName: "kube-api-access-kf2x6") pod "afc8bbf3-0297-4e5a-ba30-a8fac38c3832" (UID: "afc8bbf3-0297-4e5a-ba30-a8fac38c3832"). InnerVolumeSpecName "kube-api-access-kf2x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.644275 4778 scope.go:117] "RemoveContainer" containerID="e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e" Mar 12 14:26:10 crc kubenswrapper[4778]: E0312 14:26:10.644708 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e\": container with ID starting with e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e not found: ID does not exist" containerID="e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.644794 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e"} err="failed to get container status \"e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e\": rpc error: code = NotFound desc = could not find container \"e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e\": container with ID starting with e2fa1773eb830885c4e15af7463b9539f69e5b812f2d2bd39e3dbf38bfa0c19e not found: ID does not exist" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.644814 4778 scope.go:117] "RemoveContainer" containerID="46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b" Mar 12 14:26:10 crc kubenswrapper[4778]: E0312 14:26:10.645279 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b\": container with ID starting with 46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b not found: ID does not exist" containerID="46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.645299 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b"} err="failed to get container status \"46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b\": rpc error: code = NotFound desc = could not find container \"46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b\": container with ID starting with 46e87b0f62347e625a15fc108a812fa5374c31ccff16610cc6ccaccbc880dc3b not found: ID does not exist" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.645312 4778 scope.go:117] "RemoveContainer" containerID="72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380" Mar 12 14:26:10 crc kubenswrapper[4778]: E0312 14:26:10.645659 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380\": container with ID starting with 72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380 not found: ID does not exist" containerID="72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.645678 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380"} err="failed to get container status \"72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380\": rpc error: code = NotFound desc = could not find container \"72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380\": container with ID starting with 72109e62fc84d8faf8627ad866064d05143b1f8c544f417bf67c5efdac624380 not found: ID does not exist" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.677981 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afc8bbf3-0297-4e5a-ba30-a8fac38c3832" (UID: "afc8bbf3-0297-4e5a-ba30-a8fac38c3832"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.722954 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.722992 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:10 crc kubenswrapper[4778]: I0312 14:26:10.723003 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf2x6\" (UniqueName: \"kubernetes.io/projected/afc8bbf3-0297-4e5a-ba30-a8fac38c3832-kube-api-access-kf2x6\") on node \"crc\" DevicePath \"\"" Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.341909 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5k5f" event={"ID":"afc8bbf3-0297-4e5a-ba30-a8fac38c3832","Type":"ContainerDied","Data":"32a087503db88c25852e194ce746844351fc2c42e470484a64156247f43bc6ee"} Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.342271 4778 scope.go:117] "RemoveContainer" containerID="7996f5c983dd5f46d694007de45d6212a50151bb7f287388b92acc9fa36b446c" Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.341953 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5k5f" Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.368216 4778 scope.go:117] "RemoveContainer" containerID="089be0f6a7588cf95d5a557ff5fcbb10a29194745bfdb53617ac1aedce8267e2" Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.383045 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5k5f"] Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.396732 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d5k5f"] Mar 12 14:26:11 crc kubenswrapper[4778]: I0312 14:26:11.399617 4778 scope.go:117] "RemoveContainer" containerID="9049643c60cdd1df5b7a09b9be7298790ce9d88b86df75d5b1bc6953a508f058" Mar 12 14:26:12 crc kubenswrapper[4778]: I0312 14:26:12.278866 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" path="/var/lib/kubelet/pods/21759d07-307f-4331-94fd-5e4720ef2b7f/volumes" Mar 12 14:26:12 crc kubenswrapper[4778]: I0312 14:26:12.279912 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" path="/var/lib/kubelet/pods/afc8bbf3-0297-4e5a-ba30-a8fac38c3832/volumes" Mar 12 14:26:17 crc kubenswrapper[4778]: I0312 14:26:17.254235 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:26:17 crc kubenswrapper[4778]: E0312 14:26:17.254988 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:26:22 crc kubenswrapper[4778]: I0312 14:26:22.319378 4778 scope.go:117] "RemoveContainer" containerID="45d8e22a4c6b9a2b198c09597b6bc6f24b127ce1a5abca778cee677c28671528" Mar 12 14:26:28 crc kubenswrapper[4778]: I0312 14:26:28.253821 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:26:28 crc kubenswrapper[4778]: E0312 14:26:28.254621 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:26:39 crc kubenswrapper[4778]: I0312 14:26:39.254467 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:26:39 crc kubenswrapper[4778]: E0312 14:26:39.255169 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:26:40 crc kubenswrapper[4778]: I0312 14:26:40.527171 4778 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod21759d07-307f-4331-94fd-5e4720ef2b7f"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod21759d07-307f-4331-94fd-5e4720ef2b7f] : Timed out while waiting for systemd to remove kubepods-burstable-pod21759d07_307f_4331_94fd_5e4720ef2b7f.slice" Mar 12 14:26:52 crc kubenswrapper[4778]: I0312 14:26:52.253865 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:26:52 crc kubenswrapper[4778]: E0312 14:26:52.254801 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:27:07 crc kubenswrapper[4778]: I0312 14:27:07.253776 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:27:07 crc kubenswrapper[4778]: E0312 14:27:07.254647 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:27:22 crc kubenswrapper[4778]: I0312 14:27:22.260415 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:27:22 crc kubenswrapper[4778]: E0312 14:27:22.261223 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:27:33 crc kubenswrapper[4778]: I0312 14:27:33.254178 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:27:34 crc kubenswrapper[4778]: I0312 14:27:34.082882 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"6a1ec993be9e8a6473b90c3546089f31fa9bb55d6d9459c21a9b96e0f0006f55"} Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.145654 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555428-88vs6"] Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146761 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="registry-server" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146781 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="registry-server" Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146802 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="extract-utilities" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146811 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="extract-utilities" Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146822 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="registry-server" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146831 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="registry-server" Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146852 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="extract-content" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146861 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="extract-content" Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146878 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="extract-content" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146884 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="extract-content" Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146901 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="extract-utilities" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146908 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="extract-utilities" Mar 12 14:28:00 crc kubenswrapper[4778]: E0312 14:28:00.146925 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa45db55-92a4-4a16-9455-ee110dc34fa6" containerName="oc" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.146936 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa45db55-92a4-4a16-9455-ee110dc34fa6" containerName="oc" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.147235 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="21759d07-307f-4331-94fd-5e4720ef2b7f" containerName="registry-server" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.147254 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc8bbf3-0297-4e5a-ba30-a8fac38c3832" containerName="registry-server" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.147275 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa45db55-92a4-4a16-9455-ee110dc34fa6" containerName="oc" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.148021 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.150075 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.150393 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.157868 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.160292 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555428-88vs6"] Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.258523 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm79s\" (UniqueName: \"kubernetes.io/projected/071affdc-5584-4c5b-bfc8-f4e23c328d71-kube-api-access-jm79s\") pod \"auto-csr-approver-29555428-88vs6\" (UID: \"071affdc-5584-4c5b-bfc8-f4e23c328d71\") " pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.360537 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm79s\" (UniqueName: \"kubernetes.io/projected/071affdc-5584-4c5b-bfc8-f4e23c328d71-kube-api-access-jm79s\") pod \"auto-csr-approver-29555428-88vs6\" (UID: \"071affdc-5584-4c5b-bfc8-f4e23c328d71\") " pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.388396 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm79s\" (UniqueName: \"kubernetes.io/projected/071affdc-5584-4c5b-bfc8-f4e23c328d71-kube-api-access-jm79s\") pod \"auto-csr-approver-29555428-88vs6\" (UID: \"071affdc-5584-4c5b-bfc8-f4e23c328d71\") " pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.471277 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:00 crc kubenswrapper[4778]: I0312 14:28:00.924234 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555428-88vs6"] Mar 12 14:28:01 crc kubenswrapper[4778]: I0312 14:28:01.349179 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555428-88vs6" event={"ID":"071affdc-5584-4c5b-bfc8-f4e23c328d71","Type":"ContainerStarted","Data":"8c5ed4a7613ef08f2ef273dc42fb87d76cc0502f91ca97ea93b219c2b5f0b8da"} Mar 12 14:28:02 crc kubenswrapper[4778]: I0312 14:28:02.359236 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555428-88vs6" event={"ID":"071affdc-5584-4c5b-bfc8-f4e23c328d71","Type":"ContainerStarted","Data":"077f3532831f67d79d381fa8fb2af0e2022a1d4bb5091cf24d239727a2077516"} Mar 12 14:28:02 crc kubenswrapper[4778]: I0312 14:28:02.382002 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555428-88vs6" podStartSLOduration=1.366560612 podStartE2EDuration="2.381973524s" podCreationTimestamp="2026-03-12 14:28:00 +0000 UTC" firstStartedPulling="2026-03-12 14:28:00.936732708 +0000 UTC m=+4699.385428104" lastFinishedPulling="2026-03-12 14:28:01.95214562 +0000 UTC m=+4700.400841016" observedRunningTime="2026-03-12 14:28:02.37796651 +0000 UTC m=+4700.826661916" watchObservedRunningTime="2026-03-12 14:28:02.381973524 +0000 UTC m=+4700.830668940" Mar 12 14:28:03 crc kubenswrapper[4778]: I0312 14:28:03.369457 4778 generic.go:334] "Generic (PLEG): container finished" podID="071affdc-5584-4c5b-bfc8-f4e23c328d71" containerID="077f3532831f67d79d381fa8fb2af0e2022a1d4bb5091cf24d239727a2077516" exitCode=0 Mar 12 14:28:03 crc kubenswrapper[4778]: I0312 14:28:03.369519 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555428-88vs6" event={"ID":"071affdc-5584-4c5b-bfc8-f4e23c328d71","Type":"ContainerDied","Data":"077f3532831f67d79d381fa8fb2af0e2022a1d4bb5091cf24d239727a2077516"} Mar 12 14:28:04 crc kubenswrapper[4778]: I0312 14:28:04.845810 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:04 crc kubenswrapper[4778]: I0312 14:28:04.956055 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm79s\" (UniqueName: \"kubernetes.io/projected/071affdc-5584-4c5b-bfc8-f4e23c328d71-kube-api-access-jm79s\") pod \"071affdc-5584-4c5b-bfc8-f4e23c328d71\" (UID: \"071affdc-5584-4c5b-bfc8-f4e23c328d71\") " Mar 12 14:28:04 crc kubenswrapper[4778]: I0312 14:28:04.963360 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/071affdc-5584-4c5b-bfc8-f4e23c328d71-kube-api-access-jm79s" (OuterVolumeSpecName: "kube-api-access-jm79s") pod "071affdc-5584-4c5b-bfc8-f4e23c328d71" (UID: "071affdc-5584-4c5b-bfc8-f4e23c328d71"). InnerVolumeSpecName "kube-api-access-jm79s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:28:05 crc kubenswrapper[4778]: I0312 14:28:05.060163 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm79s\" (UniqueName: \"kubernetes.io/projected/071affdc-5584-4c5b-bfc8-f4e23c328d71-kube-api-access-jm79s\") on node \"crc\" DevicePath \"\"" Mar 12 14:28:05 crc kubenswrapper[4778]: I0312 14:28:05.386494 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555422-kw6c8"] Mar 12 14:28:05 crc kubenswrapper[4778]: I0312 14:28:05.389947 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555428-88vs6" event={"ID":"071affdc-5584-4c5b-bfc8-f4e23c328d71","Type":"ContainerDied","Data":"8c5ed4a7613ef08f2ef273dc42fb87d76cc0502f91ca97ea93b219c2b5f0b8da"} Mar 12 14:28:05 crc kubenswrapper[4778]: I0312 14:28:05.389991 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c5ed4a7613ef08f2ef273dc42fb87d76cc0502f91ca97ea93b219c2b5f0b8da" Mar 12 14:28:05 crc kubenswrapper[4778]: I0312 14:28:05.390059 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555428-88vs6" Mar 12 14:28:05 crc kubenswrapper[4778]: I0312 14:28:05.396695 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555422-kw6c8"] Mar 12 14:28:06 crc kubenswrapper[4778]: I0312 14:28:06.267019 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecab0458-0ee6-4672-bd27-4c8aae8427bb" path="/var/lib/kubelet/pods/ecab0458-0ee6-4672-bd27-4c8aae8427bb/volumes" Mar 12 14:28:22 crc kubenswrapper[4778]: I0312 14:28:22.434951 4778 scope.go:117] "RemoveContainer" containerID="460cb8eb02f9333998d559fe47fe50a7beb133708302defa156052aac3033d0e" Mar 12 14:29:58 crc kubenswrapper[4778]: I0312 14:29:58.557905 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:29:58 crc kubenswrapper[4778]: I0312 14:29:58.559545 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.168852 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl"] Mar 12 14:30:00 crc kubenswrapper[4778]: E0312 14:30:00.169821 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="071affdc-5584-4c5b-bfc8-f4e23c328d71" containerName="oc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.169838 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="071affdc-5584-4c5b-bfc8-f4e23c328d71" containerName="oc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.170063 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="071affdc-5584-4c5b-bfc8-f4e23c328d71" containerName="oc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.170946 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.173123 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.173578 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.180401 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555430-hqmdc"] Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.182023 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.184593 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.184772 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.184920 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.192357 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555430-hqmdc"] Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.203641 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl"] Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.290606 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjt9v\" (UniqueName: \"kubernetes.io/projected/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-kube-api-access-hjt9v\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.290715 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-config-volume\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.290759 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-secret-volume\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.290850 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjpnj\" (UniqueName: \"kubernetes.io/projected/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0-kube-api-access-rjpnj\") pod \"auto-csr-approver-29555430-hqmdc\" (UID: \"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0\") " pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.393071 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjt9v\" (UniqueName: \"kubernetes.io/projected/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-kube-api-access-hjt9v\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.393169 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-config-volume\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.393229 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-secret-volume\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.393318 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjpnj\" (UniqueName: \"kubernetes.io/projected/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0-kube-api-access-rjpnj\") pod \"auto-csr-approver-29555430-hqmdc\" (UID: \"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0\") " pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.394647 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-config-volume\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.399975 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-secret-volume\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.413171 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjpnj\" (UniqueName: \"kubernetes.io/projected/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0-kube-api-access-rjpnj\") pod \"auto-csr-approver-29555430-hqmdc\" (UID: \"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0\") " pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.414144 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjt9v\" (UniqueName: \"kubernetes.io/projected/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-kube-api-access-hjt9v\") pod \"collect-profiles-29555430-zhqfl\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.511358 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:00 crc kubenswrapper[4778]: I0312 14:30:00.526309 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:01 crc kubenswrapper[4778]: I0312 14:30:01.003421 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl"] Mar 12 14:30:01 crc kubenswrapper[4778]: I0312 14:30:01.012763 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555430-hqmdc"] Mar 12 14:30:01 crc kubenswrapper[4778]: I0312 14:30:01.461077 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" event={"ID":"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0","Type":"ContainerStarted","Data":"3394de743c569ccb49b14f1347f2a2010456b09cd4047ab06cda916362b6f570"} Mar 12 14:30:01 crc kubenswrapper[4778]: I0312 14:30:01.463562 4778 generic.go:334] "Generic (PLEG): container finished" podID="db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" containerID="57a58448ac2691d1255487422cd2ce72ba1abcb298bf6c4ed12464fdb32a532d" exitCode=0 Mar 12 14:30:01 crc kubenswrapper[4778]: I0312 14:30:01.463592 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" event={"ID":"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df","Type":"ContainerDied","Data":"57a58448ac2691d1255487422cd2ce72ba1abcb298bf6c4ed12464fdb32a532d"} Mar 12 14:30:01 crc kubenswrapper[4778]: I0312 14:30:01.463614 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" event={"ID":"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df","Type":"ContainerStarted","Data":"5eb849e52a7c5c3d7b24d0990176f3f994944c8351988bfe9f8003384aceae71"} Mar 12 14:30:02 crc kubenswrapper[4778]: I0312 14:30:02.477113 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" event={"ID":"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0","Type":"ContainerStarted","Data":"202ec48bafe3f3236d534da77819d177d9a1fed914f316b780eda08b2d9dcd5e"} Mar 12 14:30:02 crc kubenswrapper[4778]: I0312 14:30:02.491206 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" podStartSLOduration=1.3572657559999999 podStartE2EDuration="2.491174366s" podCreationTimestamp="2026-03-12 14:30:00 +0000 UTC" firstStartedPulling="2026-03-12 14:30:01.018942752 +0000 UTC m=+4819.467638148" lastFinishedPulling="2026-03-12 14:30:02.152851362 +0000 UTC m=+4820.601546758" observedRunningTime="2026-03-12 14:30:02.490128446 +0000 UTC m=+4820.938823852" watchObservedRunningTime="2026-03-12 14:30:02.491174366 +0000 UTC m=+4820.939869762" Mar 12 14:30:02 crc kubenswrapper[4778]: I0312 14:30:02.990415 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.150148 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-config-volume\") pod \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.150345 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjt9v\" (UniqueName: \"kubernetes.io/projected/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-kube-api-access-hjt9v\") pod \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.150407 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-secret-volume\") pod \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\" (UID: \"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df\") " Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.151138 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-config-volume" (OuterVolumeSpecName: "config-volume") pod "db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" (UID: "db4d57b8-99e5-4955-a6fe-9b0c0a6e61df"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.155961 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-kube-api-access-hjt9v" (OuterVolumeSpecName: "kube-api-access-hjt9v") pod "db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" (UID: "db4d57b8-99e5-4955-a6fe-9b0c0a6e61df"). InnerVolumeSpecName "kube-api-access-hjt9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.168487 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" (UID: "db4d57b8-99e5-4955-a6fe-9b0c0a6e61df"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.253062 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjt9v\" (UniqueName: \"kubernetes.io/projected/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-kube-api-access-hjt9v\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.253109 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.253123 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.487913 4778 generic.go:334] "Generic (PLEG): container finished" podID="e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0" containerID="202ec48bafe3f3236d534da77819d177d9a1fed914f316b780eda08b2d9dcd5e" exitCode=0 Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.488018 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" event={"ID":"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0","Type":"ContainerDied","Data":"202ec48bafe3f3236d534da77819d177d9a1fed914f316b780eda08b2d9dcd5e"} Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.490161 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" event={"ID":"db4d57b8-99e5-4955-a6fe-9b0c0a6e61df","Type":"ContainerDied","Data":"5eb849e52a7c5c3d7b24d0990176f3f994944c8351988bfe9f8003384aceae71"} Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.490303 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb849e52a7c5c3d7b24d0990176f3f994944c8351988bfe9f8003384aceae71" Mar 12 14:30:03 crc kubenswrapper[4778]: I0312 14:30:03.490219 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl" Mar 12 14:30:04 crc kubenswrapper[4778]: I0312 14:30:04.086083 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh"] Mar 12 14:30:04 crc kubenswrapper[4778]: I0312 14:30:04.099492 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555385-qwzwh"] Mar 12 14:30:04 crc kubenswrapper[4778]: I0312 14:30:04.268342 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76005d52-2d02-4a1e-89dd-c050a66fe667" path="/var/lib/kubelet/pods/76005d52-2d02-4a1e-89dd-c050a66fe667/volumes" Mar 12 14:30:04 crc kubenswrapper[4778]: I0312 14:30:04.895379 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:04 crc kubenswrapper[4778]: I0312 14:30:04.990473 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjpnj\" (UniqueName: \"kubernetes.io/projected/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0-kube-api-access-rjpnj\") pod \"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0\" (UID: \"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0\") " Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.000930 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0-kube-api-access-rjpnj" (OuterVolumeSpecName: "kube-api-access-rjpnj") pod "e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0" (UID: "e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0"). InnerVolumeSpecName "kube-api-access-rjpnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.092294 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjpnj\" (UniqueName: \"kubernetes.io/projected/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0-kube-api-access-rjpnj\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.339857 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555424-5hp4x"] Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.349044 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555424-5hp4x"] Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.509426 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" event={"ID":"e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0","Type":"ContainerDied","Data":"3394de743c569ccb49b14f1347f2a2010456b09cd4047ab06cda916362b6f570"} Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.509632 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3394de743c569ccb49b14f1347f2a2010456b09cd4047ab06cda916362b6f570" Mar 12 14:30:05 crc kubenswrapper[4778]: I0312 14:30:05.509734 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555430-hqmdc" Mar 12 14:30:06 crc kubenswrapper[4778]: I0312 14:30:06.265985 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c94941-b604-4fc0-b7b3-0b6749bbf233" path="/var/lib/kubelet/pods/78c94941-b604-4fc0-b7b3-0b6749bbf233/volumes" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.512984 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j2264"] Mar 12 14:30:20 crc kubenswrapper[4778]: E0312 14:30:20.513874 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" containerName="collect-profiles" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.513889 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" containerName="collect-profiles" Mar 12 14:30:20 crc kubenswrapper[4778]: E0312 14:30:20.513918 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0" containerName="oc" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.513925 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0" containerName="oc" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.514165 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0" containerName="oc" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.514203 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" containerName="collect-profiles" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.515796 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.541164 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2264"] Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.549822 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-utilities\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.549990 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-catalog-content\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.550050 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68l55\" (UniqueName: \"kubernetes.io/projected/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-kube-api-access-68l55\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.652970 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-utilities\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.653302 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-catalog-content\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.653630 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-utilities\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.653684 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-catalog-content\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.653788 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68l55\" (UniqueName: \"kubernetes.io/projected/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-kube-api-access-68l55\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.682544 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68l55\" (UniqueName: \"kubernetes.io/projected/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-kube-api-access-68l55\") pod \"redhat-marketplace-j2264\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:20 crc kubenswrapper[4778]: I0312 14:30:20.844699 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:21 crc kubenswrapper[4778]: I0312 14:30:21.336317 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2264"] Mar 12 14:30:21 crc kubenswrapper[4778]: I0312 14:30:21.658953 4778 generic.go:334] "Generic (PLEG): container finished" podID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerID="3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1" exitCode=0 Mar 12 14:30:21 crc kubenswrapper[4778]: I0312 14:30:21.659016 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerDied","Data":"3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1"} Mar 12 14:30:21 crc kubenswrapper[4778]: I0312 14:30:21.659320 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerStarted","Data":"746398f781b6a4873ac0e7ce0ef1a3ad903daadcf6e96df72afd204e8d36367d"} Mar 12 14:30:22 crc kubenswrapper[4778]: I0312 14:30:22.541705 4778 scope.go:117] "RemoveContainer" containerID="501e76905e9d2ff1f1e87040184d63ca0f219b530ef232d95f1fa4250e5ab145" Mar 12 14:30:22 crc kubenswrapper[4778]: I0312 14:30:22.574015 4778 scope.go:117] "RemoveContainer" containerID="a1c2cc27e654689e4f136031bb0129f78011ff7542f974149c494096b483a2a2" Mar 12 14:30:23 crc kubenswrapper[4778]: I0312 14:30:23.687797 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerStarted","Data":"ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba"} Mar 12 14:30:24 crc kubenswrapper[4778]: I0312 14:30:24.697061 4778 generic.go:334] "Generic (PLEG): container finished" podID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerID="ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba" exitCode=0 Mar 12 14:30:24 crc kubenswrapper[4778]: I0312 14:30:24.697130 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerDied","Data":"ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba"} Mar 12 14:30:25 crc kubenswrapper[4778]: I0312 14:30:25.716910 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerStarted","Data":"d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75"} Mar 12 14:30:25 crc kubenswrapper[4778]: I0312 14:30:25.755774 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j2264" podStartSLOduration=2.298192935 podStartE2EDuration="5.755751432s" podCreationTimestamp="2026-03-12 14:30:20 +0000 UTC" firstStartedPulling="2026-03-12 14:30:21.660988198 +0000 UTC m=+4840.109683594" lastFinishedPulling="2026-03-12 14:30:25.118546695 +0000 UTC m=+4843.567242091" observedRunningTime="2026-03-12 14:30:25.74546907 +0000 UTC m=+4844.194164486" watchObservedRunningTime="2026-03-12 14:30:25.755751432 +0000 UTC m=+4844.204446828" Mar 12 14:30:28 crc kubenswrapper[4778]: I0312 14:30:28.557878 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:30:28 crc kubenswrapper[4778]: I0312 14:30:28.557945 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:30:30 crc kubenswrapper[4778]: I0312 14:30:30.845397 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:30 crc kubenswrapper[4778]: I0312 14:30:30.846842 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:30 crc kubenswrapper[4778]: I0312 14:30:30.903027 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:31 crc kubenswrapper[4778]: I0312 14:30:31.846619 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:31 crc kubenswrapper[4778]: I0312 14:30:31.901598 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2264"] Mar 12 14:30:33 crc kubenswrapper[4778]: I0312 14:30:33.818288 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j2264" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="registry-server" containerID="cri-o://d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75" gracePeriod=2 Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.460580 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.558088 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-utilities\") pod \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.558139 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-catalog-content\") pod \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.558261 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68l55\" (UniqueName: \"kubernetes.io/projected/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-kube-api-access-68l55\") pod \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\" (UID: \"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac\") " Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.558874 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-utilities" (OuterVolumeSpecName: "utilities") pod "90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" (UID: "90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.565280 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-kube-api-access-68l55" (OuterVolumeSpecName: "kube-api-access-68l55") pod "90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" (UID: "90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac"). InnerVolumeSpecName "kube-api-access-68l55". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.584199 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" (UID: "90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.662032 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.662102 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.662117 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68l55\" (UniqueName: \"kubernetes.io/projected/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac-kube-api-access-68l55\") on node \"crc\" DevicePath \"\"" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.826649 4778 generic.go:334] "Generic (PLEG): container finished" podID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerID="d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75" exitCode=0 Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.826693 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerDied","Data":"d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75"} Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.826726 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2264" event={"ID":"90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac","Type":"ContainerDied","Data":"746398f781b6a4873ac0e7ce0ef1a3ad903daadcf6e96df72afd204e8d36367d"} Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.826742 4778 scope.go:117] "RemoveContainer" containerID="d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.826863 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2264" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.875316 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2264"] Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.876900 4778 scope.go:117] "RemoveContainer" containerID="ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.893024 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2264"] Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.898309 4778 scope.go:117] "RemoveContainer" containerID="3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.943457 4778 scope.go:117] "RemoveContainer" containerID="d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75" Mar 12 14:30:34 crc kubenswrapper[4778]: E0312 14:30:34.945404 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75\": container with ID starting with d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75 not found: ID does not exist" containerID="d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.945444 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75"} err="failed to get container status \"d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75\": rpc error: code = NotFound desc = could not find container \"d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75\": container with ID starting with d3cb9f2ba446c850b76c89a7cd15f986deb091a3c6401ff8166b76a3b53b0e75 not found: ID does not exist" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.945470 4778 scope.go:117] "RemoveContainer" containerID="ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba" Mar 12 14:30:34 crc kubenswrapper[4778]: E0312 14:30:34.945765 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba\": container with ID starting with ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba not found: ID does not exist" containerID="ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.945787 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba"} err="failed to get container status \"ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba\": rpc error: code = NotFound desc = could not find container \"ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba\": container with ID starting with ac315bf62854bec40ed11865b47b388a0f0bfc385b3f9ee55b4d6a4a7c7fdfba not found: ID does not exist" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.945799 4778 scope.go:117] "RemoveContainer" containerID="3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1" Mar 12 14:30:34 crc kubenswrapper[4778]: E0312 14:30:34.946069 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1\": container with ID starting with 3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1 not found: ID does not exist" containerID="3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1" Mar 12 14:30:34 crc kubenswrapper[4778]: I0312 14:30:34.946091 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1"} err="failed to get container status \"3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1\": rpc error: code = NotFound desc = could not find container \"3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1\": container with ID starting with 3d6cd7c02a9fc8f9065eb1f58c0a4587d1fd25e46f09d296a3eb1bf674d206c1 not found: ID does not exist" Mar 12 14:30:36 crc kubenswrapper[4778]: I0312 14:30:36.267390 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" path="/var/lib/kubelet/pods/90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac/volumes" Mar 12 14:30:58 crc kubenswrapper[4778]: I0312 14:30:58.558354 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:30:58 crc kubenswrapper[4778]: I0312 14:30:58.558979 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:30:58 crc kubenswrapper[4778]: I0312 14:30:58.559038 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:30:58 crc kubenswrapper[4778]: I0312 14:30:58.559935 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a1ec993be9e8a6473b90c3546089f31fa9bb55d6d9459c21a9b96e0f0006f55"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:30:58 crc kubenswrapper[4778]: I0312 14:30:58.560005 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://6a1ec993be9e8a6473b90c3546089f31fa9bb55d6d9459c21a9b96e0f0006f55" gracePeriod=600 Mar 12 14:30:59 crc kubenswrapper[4778]: I0312 14:30:59.096854 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="6a1ec993be9e8a6473b90c3546089f31fa9bb55d6d9459c21a9b96e0f0006f55" exitCode=0 Mar 12 14:30:59 crc kubenswrapper[4778]: I0312 14:30:59.097367 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"6a1ec993be9e8a6473b90c3546089f31fa9bb55d6d9459c21a9b96e0f0006f55"} Mar 12 14:30:59 crc kubenswrapper[4778]: I0312 14:30:59.097439 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c"} Mar 12 14:30:59 crc kubenswrapper[4778]: I0312 14:30:59.097462 4778 scope.go:117] "RemoveContainer" containerID="994fa1e1cf0527d97bf647f1d2a50ed301bda64c2a862df7b100daec9859483a" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.324398 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wrns6"] Mar 12 14:31:32 crc kubenswrapper[4778]: E0312 14:31:32.326408 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="extract-content" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.326432 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="extract-content" Mar 12 14:31:32 crc kubenswrapper[4778]: E0312 14:31:32.326458 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="registry-server" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.326465 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="registry-server" Mar 12 14:31:32 crc kubenswrapper[4778]: E0312 14:31:32.326477 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="extract-utilities" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.326485 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="extract-utilities" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.327109 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d6add3-4eb8-45f0-afa1-cfca4fb1e6ac" containerName="registry-server" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.329843 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.332106 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wrns6"] Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.436502 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-catalog-content\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.436563 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4969\" (UniqueName: \"kubernetes.io/projected/75d88a85-f027-434a-9c11-e5c40cb64d16-kube-api-access-c4969\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.436606 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-utilities\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.537869 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-catalog-content\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.538368 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4969\" (UniqueName: \"kubernetes.io/projected/75d88a85-f027-434a-9c11-e5c40cb64d16-kube-api-access-c4969\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.538438 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-utilities\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.538437 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-catalog-content\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.538674 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-utilities\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.565077 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4969\" (UniqueName: \"kubernetes.io/projected/75d88a85-f027-434a-9c11-e5c40cb64d16-kube-api-access-c4969\") pod \"redhat-operators-wrns6\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:32 crc kubenswrapper[4778]: I0312 14:31:32.694411 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:33 crc kubenswrapper[4778]: I0312 14:31:33.129886 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wrns6"] Mar 12 14:31:33 crc kubenswrapper[4778]: I0312 14:31:33.418998 4778 generic.go:334] "Generic (PLEG): container finished" podID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerID="c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538" exitCode=0 Mar 12 14:31:33 crc kubenswrapper[4778]: I0312 14:31:33.419094 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerDied","Data":"c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538"} Mar 12 14:31:33 crc kubenswrapper[4778]: I0312 14:31:33.419350 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerStarted","Data":"1c7c3eb15ff3c903a138dcd8c30e7f50b560dc5fe6819c4ba964d113870db2f2"} Mar 12 14:31:33 crc kubenswrapper[4778]: I0312 14:31:33.420843 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:31:34 crc kubenswrapper[4778]: I0312 14:31:34.430601 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerStarted","Data":"8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351"} Mar 12 14:31:39 crc kubenswrapper[4778]: I0312 14:31:39.479214 4778 generic.go:334] "Generic (PLEG): container finished" podID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerID="8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351" exitCode=0 Mar 12 14:31:39 crc kubenswrapper[4778]: I0312 14:31:39.479306 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerDied","Data":"8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351"} Mar 12 14:31:40 crc kubenswrapper[4778]: I0312 14:31:40.494308 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerStarted","Data":"d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54"} Mar 12 14:31:40 crc kubenswrapper[4778]: I0312 14:31:40.525122 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wrns6" podStartSLOduration=1.9610731750000001 podStartE2EDuration="8.525097252s" podCreationTimestamp="2026-03-12 14:31:32 +0000 UTC" firstStartedPulling="2026-03-12 14:31:33.420611627 +0000 UTC m=+4911.869307023" lastFinishedPulling="2026-03-12 14:31:39.984635704 +0000 UTC m=+4918.433331100" observedRunningTime="2026-03-12 14:31:40.511609639 +0000 UTC m=+4918.960305045" watchObservedRunningTime="2026-03-12 14:31:40.525097252 +0000 UTC m=+4918.973792668" Mar 12 14:31:42 crc kubenswrapper[4778]: I0312 14:31:42.695159 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:42 crc kubenswrapper[4778]: I0312 14:31:42.696225 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:31:43 crc kubenswrapper[4778]: I0312 14:31:43.744106 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wrns6" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="registry-server" probeResult="failure" output=< Mar 12 14:31:43 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:31:43 crc kubenswrapper[4778]: > Mar 12 14:31:53 crc kubenswrapper[4778]: I0312 14:31:53.741263 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wrns6" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="registry-server" probeResult="failure" output=< Mar 12 14:31:53 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:31:53 crc kubenswrapper[4778]: > Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.152744 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555432-94dlz"] Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.154692 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.156463 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.156840 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.157048 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.166330 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555432-94dlz"] Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.261220 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5md2x\" (UniqueName: \"kubernetes.io/projected/4b6391c3-533c-4b44-b1be-2a5c9752ba4b-kube-api-access-5md2x\") pod \"auto-csr-approver-29555432-94dlz\" (UID: \"4b6391c3-533c-4b44-b1be-2a5c9752ba4b\") " pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.363646 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5md2x\" (UniqueName: \"kubernetes.io/projected/4b6391c3-533c-4b44-b1be-2a5c9752ba4b-kube-api-access-5md2x\") pod \"auto-csr-approver-29555432-94dlz\" (UID: \"4b6391c3-533c-4b44-b1be-2a5c9752ba4b\") " pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.383372 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5md2x\" (UniqueName: \"kubernetes.io/projected/4b6391c3-533c-4b44-b1be-2a5c9752ba4b-kube-api-access-5md2x\") pod \"auto-csr-approver-29555432-94dlz\" (UID: \"4b6391c3-533c-4b44-b1be-2a5c9752ba4b\") " pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.473873 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:00 crc kubenswrapper[4778]: I0312 14:32:00.965310 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555432-94dlz"] Mar 12 14:32:01 crc kubenswrapper[4778]: I0312 14:32:01.668479 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555432-94dlz" event={"ID":"4b6391c3-533c-4b44-b1be-2a5c9752ba4b","Type":"ContainerStarted","Data":"d4b6c9d11eb3ff84a8caa2505ac01209847a1935674926fa5a2116afc7245310"} Mar 12 14:32:02 crc kubenswrapper[4778]: I0312 14:32:02.744465 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:32:02 crc kubenswrapper[4778]: I0312 14:32:02.812727 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:32:03 crc kubenswrapper[4778]: I0312 14:32:03.510136 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wrns6"] Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:04.696291 4778 generic.go:334] "Generic (PLEG): container finished" podID="4b6391c3-533c-4b44-b1be-2a5c9752ba4b" containerID="f68c8ed6b7c2e6259023580b179d97b5ef4d89ae76842473f005cc28f0933cea" exitCode=0 Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:04.697037 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wrns6" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="registry-server" containerID="cri-o://d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54" gracePeriod=2 Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:04.697173 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555432-94dlz" event={"ID":"4b6391c3-533c-4b44-b1be-2a5c9752ba4b","Type":"ContainerDied","Data":"f68c8ed6b7c2e6259023580b179d97b5ef4d89ae76842473f005cc28f0933cea"} Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.284896 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.456439 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4969\" (UniqueName: \"kubernetes.io/projected/75d88a85-f027-434a-9c11-e5c40cb64d16-kube-api-access-c4969\") pod \"75d88a85-f027-434a-9c11-e5c40cb64d16\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.456792 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-catalog-content\") pod \"75d88a85-f027-434a-9c11-e5c40cb64d16\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.456932 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-utilities\") pod \"75d88a85-f027-434a-9c11-e5c40cb64d16\" (UID: \"75d88a85-f027-434a-9c11-e5c40cb64d16\") " Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.458530 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-utilities" (OuterVolumeSpecName: "utilities") pod "75d88a85-f027-434a-9c11-e5c40cb64d16" (UID: "75d88a85-f027-434a-9c11-e5c40cb64d16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.463260 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d88a85-f027-434a-9c11-e5c40cb64d16-kube-api-access-c4969" (OuterVolumeSpecName: "kube-api-access-c4969") pod "75d88a85-f027-434a-9c11-e5c40cb64d16" (UID: "75d88a85-f027-434a-9c11-e5c40cb64d16"). InnerVolumeSpecName "kube-api-access-c4969". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.558987 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.559022 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4969\" (UniqueName: \"kubernetes.io/projected/75d88a85-f027-434a-9c11-e5c40cb64d16-kube-api-access-c4969\") on node \"crc\" DevicePath \"\"" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.587252 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75d88a85-f027-434a-9c11-e5c40cb64d16" (UID: "75d88a85-f027-434a-9c11-e5c40cb64d16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.660766 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d88a85-f027-434a-9c11-e5c40cb64d16-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.719870 4778 generic.go:334] "Generic (PLEG): container finished" podID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerID="d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54" exitCode=0 Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.720062 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wrns6" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.722841 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerDied","Data":"d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54"} Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.722898 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wrns6" event={"ID":"75d88a85-f027-434a-9c11-e5c40cb64d16","Type":"ContainerDied","Data":"1c7c3eb15ff3c903a138dcd8c30e7f50b560dc5fe6819c4ba964d113870db2f2"} Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.722927 4778 scope.go:117] "RemoveContainer" containerID="d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54" Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.762247 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wrns6"] Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.771258 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wrns6"] Mar 12 14:32:05 crc kubenswrapper[4778]: I0312 14:32:05.773227 4778 scope.go:117] "RemoveContainer" containerID="8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.154011 4778 scope.go:117] "RemoveContainer" containerID="c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.236120 4778 scope.go:117] "RemoveContainer" containerID="d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54" Mar 12 14:32:06 crc kubenswrapper[4778]: E0312 14:32:06.236678 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54\": container with ID starting with d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54 not found: ID does not exist" containerID="d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.236747 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54"} err="failed to get container status \"d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54\": rpc error: code = NotFound desc = could not find container \"d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54\": container with ID starting with d303c9bc7a74d64545141c564ba3f0be945f88b7f92e9057d960eb47fc781a54 not found: ID does not exist" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.236788 4778 scope.go:117] "RemoveContainer" containerID="8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351" Mar 12 14:32:06 crc kubenswrapper[4778]: E0312 14:32:06.237139 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351\": container with ID starting with 8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351 not found: ID does not exist" containerID="8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.237176 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351"} err="failed to get container status \"8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351\": rpc error: code = NotFound desc = could not find container \"8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351\": container with ID starting with 8db1f0fdafeaee661df2b6ada567ab3179ef90cc4d4e57d242e9a091b1bfe351 not found: ID does not exist" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.237209 4778 scope.go:117] "RemoveContainer" containerID="c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538" Mar 12 14:32:06 crc kubenswrapper[4778]: E0312 14:32:06.237446 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538\": container with ID starting with c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538 not found: ID does not exist" containerID="c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.237473 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538"} err="failed to get container status \"c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538\": rpc error: code = NotFound desc = could not find container \"c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538\": container with ID starting with c68406c6882aec720e9611447e1ccc8988ea75684e147b1609224dcddd31d538 not found: ID does not exist" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.264495 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" path="/var/lib/kubelet/pods/75d88a85-f027-434a-9c11-e5c40cb64d16/volumes" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.276419 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.374576 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5md2x\" (UniqueName: \"kubernetes.io/projected/4b6391c3-533c-4b44-b1be-2a5c9752ba4b-kube-api-access-5md2x\") pod \"4b6391c3-533c-4b44-b1be-2a5c9752ba4b\" (UID: \"4b6391c3-533c-4b44-b1be-2a5c9752ba4b\") " Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.379521 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b6391c3-533c-4b44-b1be-2a5c9752ba4b-kube-api-access-5md2x" (OuterVolumeSpecName: "kube-api-access-5md2x") pod "4b6391c3-533c-4b44-b1be-2a5c9752ba4b" (UID: "4b6391c3-533c-4b44-b1be-2a5c9752ba4b"). InnerVolumeSpecName "kube-api-access-5md2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.477890 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5md2x\" (UniqueName: \"kubernetes.io/projected/4b6391c3-533c-4b44-b1be-2a5c9752ba4b-kube-api-access-5md2x\") on node \"crc\" DevicePath \"\"" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.729933 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555432-94dlz" event={"ID":"4b6391c3-533c-4b44-b1be-2a5c9752ba4b","Type":"ContainerDied","Data":"d4b6c9d11eb3ff84a8caa2505ac01209847a1935674926fa5a2116afc7245310"} Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.730567 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4b6c9d11eb3ff84a8caa2505ac01209847a1935674926fa5a2116afc7245310" Mar 12 14:32:06 crc kubenswrapper[4778]: I0312 14:32:06.729978 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555432-94dlz" Mar 12 14:32:07 crc kubenswrapper[4778]: I0312 14:32:07.358894 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555426-btwds"] Mar 12 14:32:07 crc kubenswrapper[4778]: I0312 14:32:07.368527 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555426-btwds"] Mar 12 14:32:08 crc kubenswrapper[4778]: I0312 14:32:08.264976 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa45db55-92a4-4a16-9455-ee110dc34fa6" path="/var/lib/kubelet/pods/fa45db55-92a4-4a16-9455-ee110dc34fa6/volumes" Mar 12 14:32:22 crc kubenswrapper[4778]: I0312 14:32:22.735310 4778 scope.go:117] "RemoveContainer" containerID="188dd1cb886e6788ffd8398573fda57dc92b1fe481e6f2ffdc97a0e049e9348c" Mar 12 14:32:58 crc kubenswrapper[4778]: I0312 14:32:58.557599 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:32:58 crc kubenswrapper[4778]: I0312 14:32:58.558307 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:33:28 crc kubenswrapper[4778]: I0312 14:33:28.557333 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:33:28 crc kubenswrapper[4778]: I0312 14:33:28.557966 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.558459 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.559308 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.559397 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.560751 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.560892 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" gracePeriod=600 Mar 12 14:33:58 crc kubenswrapper[4778]: E0312 14:33:58.701486 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.721352 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" exitCode=0 Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.721393 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c"} Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.721425 4778 scope.go:117] "RemoveContainer" containerID="6a1ec993be9e8a6473b90c3546089f31fa9bb55d6d9459c21a9b96e0f0006f55" Mar 12 14:33:58 crc kubenswrapper[4778]: I0312 14:33:58.722367 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:33:58 crc kubenswrapper[4778]: E0312 14:33:58.722890 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.147870 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555434-hzddc"] Mar 12 14:34:00 crc kubenswrapper[4778]: E0312 14:34:00.149560 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="extract-utilities" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.149578 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="extract-utilities" Mar 12 14:34:00 crc kubenswrapper[4778]: E0312 14:34:00.149618 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="extract-content" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.149628 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="extract-content" Mar 12 14:34:00 crc kubenswrapper[4778]: E0312 14:34:00.149648 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="registry-server" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.149656 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="registry-server" Mar 12 14:34:00 crc kubenswrapper[4778]: E0312 14:34:00.149667 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6391c3-533c-4b44-b1be-2a5c9752ba4b" containerName="oc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.149674 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6391c3-533c-4b44-b1be-2a5c9752ba4b" containerName="oc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.149910 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b6391c3-533c-4b44-b1be-2a5c9752ba4b" containerName="oc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.149942 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d88a85-f027-434a-9c11-e5c40cb64d16" containerName="registry-server" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.150723 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.155392 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.155967 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.157992 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.181981 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555434-hzddc"] Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.272242 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6td7\" (UniqueName: \"kubernetes.io/projected/e7855ae5-9f57-4d62-ab01-d16ae9f5a037-kube-api-access-w6td7\") pod \"auto-csr-approver-29555434-hzddc\" (UID: \"e7855ae5-9f57-4d62-ab01-d16ae9f5a037\") " pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.374457 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6td7\" (UniqueName: \"kubernetes.io/projected/e7855ae5-9f57-4d62-ab01-d16ae9f5a037-kube-api-access-w6td7\") pod \"auto-csr-approver-29555434-hzddc\" (UID: \"e7855ae5-9f57-4d62-ab01-d16ae9f5a037\") " pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.397850 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6td7\" (UniqueName: \"kubernetes.io/projected/e7855ae5-9f57-4d62-ab01-d16ae9f5a037-kube-api-access-w6td7\") pod \"auto-csr-approver-29555434-hzddc\" (UID: \"e7855ae5-9f57-4d62-ab01-d16ae9f5a037\") " pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.473509 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:00 crc kubenswrapper[4778]: I0312 14:34:00.945167 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555434-hzddc"] Mar 12 14:34:01 crc kubenswrapper[4778]: I0312 14:34:01.751558 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555434-hzddc" event={"ID":"e7855ae5-9f57-4d62-ab01-d16ae9f5a037","Type":"ContainerStarted","Data":"d4fe264c15cb5db14849184e4b38afe0dbec8fa34e51673b8ea8ad0e2910d3db"} Mar 12 14:34:02 crc kubenswrapper[4778]: I0312 14:34:02.765141 4778 generic.go:334] "Generic (PLEG): container finished" podID="e7855ae5-9f57-4d62-ab01-d16ae9f5a037" containerID="ecf6cfdc210df01866b5bda8e874db3a9407a84531517ea05fb802b9d57bcdb0" exitCode=0 Mar 12 14:34:02 crc kubenswrapper[4778]: I0312 14:34:02.765210 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555434-hzddc" event={"ID":"e7855ae5-9f57-4d62-ab01-d16ae9f5a037","Type":"ContainerDied","Data":"ecf6cfdc210df01866b5bda8e874db3a9407a84531517ea05fb802b9d57bcdb0"} Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.645662 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.764274 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6td7\" (UniqueName: \"kubernetes.io/projected/e7855ae5-9f57-4d62-ab01-d16ae9f5a037-kube-api-access-w6td7\") pod \"e7855ae5-9f57-4d62-ab01-d16ae9f5a037\" (UID: \"e7855ae5-9f57-4d62-ab01-d16ae9f5a037\") " Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.770219 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7855ae5-9f57-4d62-ab01-d16ae9f5a037-kube-api-access-w6td7" (OuterVolumeSpecName: "kube-api-access-w6td7") pod "e7855ae5-9f57-4d62-ab01-d16ae9f5a037" (UID: "e7855ae5-9f57-4d62-ab01-d16ae9f5a037"). InnerVolumeSpecName "kube-api-access-w6td7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.784606 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555434-hzddc" event={"ID":"e7855ae5-9f57-4d62-ab01-d16ae9f5a037","Type":"ContainerDied","Data":"d4fe264c15cb5db14849184e4b38afe0dbec8fa34e51673b8ea8ad0e2910d3db"} Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.784650 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4fe264c15cb5db14849184e4b38afe0dbec8fa34e51673b8ea8ad0e2910d3db" Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.784724 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555434-hzddc" Mar 12 14:34:04 crc kubenswrapper[4778]: I0312 14:34:04.867542 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6td7\" (UniqueName: \"kubernetes.io/projected/e7855ae5-9f57-4d62-ab01-d16ae9f5a037-kube-api-access-w6td7\") on node \"crc\" DevicePath \"\"" Mar 12 14:34:05 crc kubenswrapper[4778]: I0312 14:34:05.723216 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555428-88vs6"] Mar 12 14:34:05 crc kubenswrapper[4778]: I0312 14:34:05.735022 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555428-88vs6"] Mar 12 14:34:06 crc kubenswrapper[4778]: I0312 14:34:06.264994 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="071affdc-5584-4c5b-bfc8-f4e23c328d71" path="/var/lib/kubelet/pods/071affdc-5584-4c5b-bfc8-f4e23c328d71/volumes" Mar 12 14:34:14 crc kubenswrapper[4778]: I0312 14:34:14.253581 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:34:14 crc kubenswrapper[4778]: E0312 14:34:14.254421 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:34:22 crc kubenswrapper[4778]: I0312 14:34:22.844694 4778 scope.go:117] "RemoveContainer" containerID="077f3532831f67d79d381fa8fb2af0e2022a1d4bb5091cf24d239727a2077516" Mar 12 14:34:26 crc kubenswrapper[4778]: I0312 14:34:26.254560 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:34:26 crc kubenswrapper[4778]: E0312 14:34:26.255387 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:34:40 crc kubenswrapper[4778]: I0312 14:34:40.256230 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:34:40 crc kubenswrapper[4778]: E0312 14:34:40.256915 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:34:52 crc kubenswrapper[4778]: I0312 14:34:52.261284 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:34:52 crc kubenswrapper[4778]: E0312 14:34:52.262071 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:35:04 crc kubenswrapper[4778]: I0312 14:35:04.254677 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:35:04 crc kubenswrapper[4778]: E0312 14:35:04.255520 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:35:15 crc kubenswrapper[4778]: I0312 14:35:15.255499 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:35:15 crc kubenswrapper[4778]: E0312 14:35:15.257752 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:35:27 crc kubenswrapper[4778]: I0312 14:35:27.254718 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:35:27 crc kubenswrapper[4778]: E0312 14:35:27.255948 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:35:38 crc kubenswrapper[4778]: I0312 14:35:38.254467 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:35:38 crc kubenswrapper[4778]: E0312 14:35:38.255124 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:35:50 crc kubenswrapper[4778]: I0312 14:35:50.254601 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:35:50 crc kubenswrapper[4778]: E0312 14:35:50.255537 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.144042 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555436-9bs4m"] Mar 12 14:36:00 crc kubenswrapper[4778]: E0312 14:36:00.145114 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7855ae5-9f57-4d62-ab01-d16ae9f5a037" containerName="oc" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.145131 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7855ae5-9f57-4d62-ab01-d16ae9f5a037" containerName="oc" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.145449 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7855ae5-9f57-4d62-ab01-d16ae9f5a037" containerName="oc" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.146220 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.148031 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.148292 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.150347 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.153773 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555436-9bs4m"] Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.280289 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwgwh\" (UniqueName: \"kubernetes.io/projected/45d1f962-b71e-4473-b387-137a395e1a39-kube-api-access-wwgwh\") pod \"auto-csr-approver-29555436-9bs4m\" (UID: \"45d1f962-b71e-4473-b387-137a395e1a39\") " pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.381951 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwgwh\" (UniqueName: \"kubernetes.io/projected/45d1f962-b71e-4473-b387-137a395e1a39-kube-api-access-wwgwh\") pod \"auto-csr-approver-29555436-9bs4m\" (UID: \"45d1f962-b71e-4473-b387-137a395e1a39\") " pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.404042 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwgwh\" (UniqueName: \"kubernetes.io/projected/45d1f962-b71e-4473-b387-137a395e1a39-kube-api-access-wwgwh\") pod \"auto-csr-approver-29555436-9bs4m\" (UID: \"45d1f962-b71e-4473-b387-137a395e1a39\") " pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.467920 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:00 crc kubenswrapper[4778]: I0312 14:36:00.910802 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555436-9bs4m"] Mar 12 14:36:01 crc kubenswrapper[4778]: W0312 14:36:01.032498 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d1f962_b71e_4473_b387_137a395e1a39.slice/crio-46c218281802b0c8f5d068642d635879a43981b837f5f8787e72274312e00084 WatchSource:0}: Error finding container 46c218281802b0c8f5d068642d635879a43981b837f5f8787e72274312e00084: Status 404 returned error can't find the container with id 46c218281802b0c8f5d068642d635879a43981b837f5f8787e72274312e00084 Mar 12 14:36:01 crc kubenswrapper[4778]: I0312 14:36:01.885436 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" event={"ID":"45d1f962-b71e-4473-b387-137a395e1a39","Type":"ContainerStarted","Data":"46c218281802b0c8f5d068642d635879a43981b837f5f8787e72274312e00084"} Mar 12 14:36:02 crc kubenswrapper[4778]: I0312 14:36:02.896758 4778 generic.go:334] "Generic (PLEG): container finished" podID="45d1f962-b71e-4473-b387-137a395e1a39" containerID="b4d039fad9b993f652c5f6f0f661f085d4f93b47467c47a0fe13959b9f367b5d" exitCode=0 Mar 12 14:36:02 crc kubenswrapper[4778]: I0312 14:36:02.896850 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" event={"ID":"45d1f962-b71e-4473-b387-137a395e1a39","Type":"ContainerDied","Data":"b4d039fad9b993f652c5f6f0f661f085d4f93b47467c47a0fe13959b9f367b5d"} Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.342941 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.468793 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwgwh\" (UniqueName: \"kubernetes.io/projected/45d1f962-b71e-4473-b387-137a395e1a39-kube-api-access-wwgwh\") pod \"45d1f962-b71e-4473-b387-137a395e1a39\" (UID: \"45d1f962-b71e-4473-b387-137a395e1a39\") " Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.477080 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d1f962-b71e-4473-b387-137a395e1a39-kube-api-access-wwgwh" (OuterVolumeSpecName: "kube-api-access-wwgwh") pod "45d1f962-b71e-4473-b387-137a395e1a39" (UID: "45d1f962-b71e-4473-b387-137a395e1a39"). InnerVolumeSpecName "kube-api-access-wwgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.571873 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwgwh\" (UniqueName: \"kubernetes.io/projected/45d1f962-b71e-4473-b387-137a395e1a39-kube-api-access-wwgwh\") on node \"crc\" DevicePath \"\"" Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.939885 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" event={"ID":"45d1f962-b71e-4473-b387-137a395e1a39","Type":"ContainerDied","Data":"46c218281802b0c8f5d068642d635879a43981b837f5f8787e72274312e00084"} Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.939936 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46c218281802b0c8f5d068642d635879a43981b837f5f8787e72274312e00084" Mar 12 14:36:04 crc kubenswrapper[4778]: I0312 14:36:04.940000 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555436-9bs4m" Mar 12 14:36:05 crc kubenswrapper[4778]: I0312 14:36:05.254262 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:36:05 crc kubenswrapper[4778]: E0312 14:36:05.254475 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:36:05 crc kubenswrapper[4778]: I0312 14:36:05.410834 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555430-hqmdc"] Mar 12 14:36:05 crc kubenswrapper[4778]: I0312 14:36:05.425369 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555430-hqmdc"] Mar 12 14:36:06 crc kubenswrapper[4778]: I0312 14:36:06.267114 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0" path="/var/lib/kubelet/pods/e89b5d9b-fc4f-4dd6-aae9-9d1ee7b3edb0/volumes" Mar 12 14:36:16 crc kubenswrapper[4778]: I0312 14:36:16.254567 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:36:16 crc kubenswrapper[4778]: E0312 14:36:16.256922 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:36:22 crc kubenswrapper[4778]: I0312 14:36:22.936109 4778 scope.go:117] "RemoveContainer" containerID="202ec48bafe3f3236d534da77819d177d9a1fed914f316b780eda08b2d9dcd5e" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.870286 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 14:36:29 crc kubenswrapper[4778]: E0312 14:36:29.871496 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d1f962-b71e-4473-b387-137a395e1a39" containerName="oc" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.871512 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d1f962-b71e-4473-b387-137a395e1a39" containerName="oc" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.871713 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d1f962-b71e-4473-b387-137a395e1a39" containerName="oc" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.873483 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.885773 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.983984 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-utilities\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.984294 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-catalog-content\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:29 crc kubenswrapper[4778]: I0312 14:36:29.984391 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2cq2\" (UniqueName: \"kubernetes.io/projected/9098edbc-6c4b-444b-8214-5848756ec94b-kube-api-access-p2cq2\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.087418 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-utilities\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.087892 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-catalog-content\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.088045 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-utilities\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.088144 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2cq2\" (UniqueName: \"kubernetes.io/projected/9098edbc-6c4b-444b-8214-5848756ec94b-kube-api-access-p2cq2\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.088511 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-catalog-content\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.110318 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2cq2\" (UniqueName: \"kubernetes.io/projected/9098edbc-6c4b-444b-8214-5848756ec94b-kube-api-access-p2cq2\") pod \"community-operators-bthl5\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.205925 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:30 crc kubenswrapper[4778]: I0312 14:36:30.730052 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 14:36:31 crc kubenswrapper[4778]: I0312 14:36:31.231770 4778 generic.go:334] "Generic (PLEG): container finished" podID="9098edbc-6c4b-444b-8214-5848756ec94b" containerID="f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682" exitCode=0 Mar 12 14:36:31 crc kubenswrapper[4778]: I0312 14:36:31.231828 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bthl5" event={"ID":"9098edbc-6c4b-444b-8214-5848756ec94b","Type":"ContainerDied","Data":"f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682"} Mar 12 14:36:31 crc kubenswrapper[4778]: I0312 14:36:31.232073 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bthl5" event={"ID":"9098edbc-6c4b-444b-8214-5848756ec94b","Type":"ContainerStarted","Data":"7e43af4c8ac9f109aea2498c7d43bec693ffd79761be06aa8860c32373c46a08"} Mar 12 14:36:31 crc kubenswrapper[4778]: I0312 14:36:31.253831 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:36:31 crc kubenswrapper[4778]: E0312 14:36:31.254080 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:36:36 crc kubenswrapper[4778]: I0312 14:36:36.304212 4778 generic.go:334] "Generic (PLEG): container finished" podID="9098edbc-6c4b-444b-8214-5848756ec94b" containerID="f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88" exitCode=0 Mar 12 14:36:36 crc kubenswrapper[4778]: I0312 14:36:36.304669 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bthl5" event={"ID":"9098edbc-6c4b-444b-8214-5848756ec94b","Type":"ContainerDied","Data":"f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88"} Mar 12 14:36:36 crc kubenswrapper[4778]: I0312 14:36:36.306525 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:36:37 crc kubenswrapper[4778]: I0312 14:36:37.316882 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bthl5" event={"ID":"9098edbc-6c4b-444b-8214-5848756ec94b","Type":"ContainerStarted","Data":"db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64"} Mar 12 14:36:37 crc kubenswrapper[4778]: I0312 14:36:37.343855 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bthl5" podStartSLOduration=2.786556618 podStartE2EDuration="8.343835425s" podCreationTimestamp="2026-03-12 14:36:29 +0000 UTC" firstStartedPulling="2026-03-12 14:36:31.233724718 +0000 UTC m=+5209.682420114" lastFinishedPulling="2026-03-12 14:36:36.791003515 +0000 UTC m=+5215.239698921" observedRunningTime="2026-03-12 14:36:37.336555938 +0000 UTC m=+5215.785251354" watchObservedRunningTime="2026-03-12 14:36:37.343835425 +0000 UTC m=+5215.792530821" Mar 12 14:36:40 crc kubenswrapper[4778]: I0312 14:36:40.206071 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:40 crc kubenswrapper[4778]: I0312 14:36:40.206416 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:40 crc kubenswrapper[4778]: I0312 14:36:40.263659 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:44 crc kubenswrapper[4778]: I0312 14:36:44.254286 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:36:44 crc kubenswrapper[4778]: E0312 14:36:44.255062 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:36:50 crc kubenswrapper[4778]: I0312 14:36:50.266697 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bthl5" Mar 12 14:36:50 crc kubenswrapper[4778]: I0312 14:36:50.352931 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 14:36:50 crc kubenswrapper[4778]: I0312 14:36:50.400511 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 14:36:50 crc kubenswrapper[4778]: I0312 14:36:50.400991 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zbbfg" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="registry-server" containerID="cri-o://ed078967e33cb44c74a365a9804f9a8509ee01d3f7a8039f9f7b8f3366ab7aae" gracePeriod=2 Mar 12 14:36:50 crc kubenswrapper[4778]: I0312 14:36:50.544454 4778 generic.go:334] "Generic (PLEG): container finished" podID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerID="ed078967e33cb44c74a365a9804f9a8509ee01d3f7a8039f9f7b8f3366ab7aae" exitCode=0 Mar 12 14:36:50 crc kubenswrapper[4778]: I0312 14:36:50.545177 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbbfg" event={"ID":"c8cc55b1-e6ed-4790-886c-fabe5917bf27","Type":"ContainerDied","Data":"ed078967e33cb44c74a365a9804f9a8509ee01d3f7a8039f9f7b8f3366ab7aae"} Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.272086 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.440510 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-catalog-content\") pod \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.440601 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d6fn\" (UniqueName: \"kubernetes.io/projected/c8cc55b1-e6ed-4790-886c-fabe5917bf27-kube-api-access-8d6fn\") pod \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.440647 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-utilities\") pod \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\" (UID: \"c8cc55b1-e6ed-4790-886c-fabe5917bf27\") " Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.445421 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-utilities" (OuterVolumeSpecName: "utilities") pod "c8cc55b1-e6ed-4790-886c-fabe5917bf27" (UID: "c8cc55b1-e6ed-4790-886c-fabe5917bf27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.448485 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8cc55b1-e6ed-4790-886c-fabe5917bf27-kube-api-access-8d6fn" (OuterVolumeSpecName: "kube-api-access-8d6fn") pod "c8cc55b1-e6ed-4790-886c-fabe5917bf27" (UID: "c8cc55b1-e6ed-4790-886c-fabe5917bf27"). InnerVolumeSpecName "kube-api-access-8d6fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.525237 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8cc55b1-e6ed-4790-886c-fabe5917bf27" (UID: "c8cc55b1-e6ed-4790-886c-fabe5917bf27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.543547 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.543577 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d6fn\" (UniqueName: \"kubernetes.io/projected/c8cc55b1-e6ed-4790-886c-fabe5917bf27-kube-api-access-8d6fn\") on node \"crc\" DevicePath \"\"" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.543589 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cc55b1-e6ed-4790-886c-fabe5917bf27-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.559442 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbbfg" event={"ID":"c8cc55b1-e6ed-4790-886c-fabe5917bf27","Type":"ContainerDied","Data":"7cfbf75bc1bea8190b4fd8a7b4f36c4f8056d3512bf0a0494d17fb32c82abce1"} Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.559493 4778 scope.go:117] "RemoveContainer" containerID="ed078967e33cb44c74a365a9804f9a8509ee01d3f7a8039f9f7b8f3366ab7aae" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.559657 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbbfg" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.594274 4778 scope.go:117] "RemoveContainer" containerID="bdf54c6d37ca16db7981b38aa8bdf481e8ce434ef1861261a6875f0a169c6607" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.618268 4778 scope.go:117] "RemoveContainer" containerID="a44a31875240c27026c8d5b3562efaf0a4ac960ee6a568ff9dac9567138bfecd" Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.642244 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 14:36:51 crc kubenswrapper[4778]: I0312 14:36:51.653952 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zbbfg"] Mar 12 14:36:52 crc kubenswrapper[4778]: I0312 14:36:52.268637 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" path="/var/lib/kubelet/pods/c8cc55b1-e6ed-4790-886c-fabe5917bf27/volumes" Mar 12 14:36:55 crc kubenswrapper[4778]: I0312 14:36:55.254088 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:36:55 crc kubenswrapper[4778]: E0312 14:36:55.255062 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.539428 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xr4w"] Mar 12 14:36:57 crc kubenswrapper[4778]: E0312 14:36:57.540256 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="extract-content" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.540273 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="extract-content" Mar 12 14:36:57 crc kubenswrapper[4778]: E0312 14:36:57.540291 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="extract-utilities" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.540300 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="extract-utilities" Mar 12 14:36:57 crc kubenswrapper[4778]: E0312 14:36:57.540340 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="registry-server" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.540349 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="registry-server" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.540568 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8cc55b1-e6ed-4790-886c-fabe5917bf27" containerName="registry-server" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.541930 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.567824 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xr4w"] Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.718135 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-catalog-content\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.718223 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-utilities\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.718269 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf5jk\" (UniqueName: \"kubernetes.io/projected/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-kube-api-access-sf5jk\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.819821 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-catalog-content\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.819894 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-utilities\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.819929 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf5jk\" (UniqueName: \"kubernetes.io/projected/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-kube-api-access-sf5jk\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.820519 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-catalog-content\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.820531 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-utilities\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.838778 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf5jk\" (UniqueName: \"kubernetes.io/projected/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-kube-api-access-sf5jk\") pod \"certified-operators-6xr4w\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:57 crc kubenswrapper[4778]: I0312 14:36:57.859904 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:36:58 crc kubenswrapper[4778]: I0312 14:36:58.452386 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xr4w"] Mar 12 14:36:58 crc kubenswrapper[4778]: I0312 14:36:58.622543 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerStarted","Data":"64e2681b11aea9607bf55a800b77d410d04225bde783d6a66b217bb9ae3cf27c"} Mar 12 14:36:59 crc kubenswrapper[4778]: I0312 14:36:59.641719 4778 generic.go:334] "Generic (PLEG): container finished" podID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerID="71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144" exitCode=0 Mar 12 14:36:59 crc kubenswrapper[4778]: I0312 14:36:59.641956 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerDied","Data":"71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144"} Mar 12 14:37:00 crc kubenswrapper[4778]: I0312 14:37:00.652647 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerStarted","Data":"3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea"} Mar 12 14:37:02 crc kubenswrapper[4778]: I0312 14:37:02.671278 4778 generic.go:334] "Generic (PLEG): container finished" podID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerID="3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea" exitCode=0 Mar 12 14:37:02 crc kubenswrapper[4778]: I0312 14:37:02.671327 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerDied","Data":"3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea"} Mar 12 14:37:04 crc kubenswrapper[4778]: I0312 14:37:04.693276 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerStarted","Data":"0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53"} Mar 12 14:37:04 crc kubenswrapper[4778]: I0312 14:37:04.757445 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xr4w" podStartSLOduration=4.242362802 podStartE2EDuration="7.757414081s" podCreationTimestamp="2026-03-12 14:36:57 +0000 UTC" firstStartedPulling="2026-03-12 14:36:59.644024474 +0000 UTC m=+5238.092719880" lastFinishedPulling="2026-03-12 14:37:03.159075763 +0000 UTC m=+5241.607771159" observedRunningTime="2026-03-12 14:37:04.708922392 +0000 UTC m=+5243.157617808" watchObservedRunningTime="2026-03-12 14:37:04.757414081 +0000 UTC m=+5243.206109507" Mar 12 14:37:07 crc kubenswrapper[4778]: I0312 14:37:07.860368 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:37:07 crc kubenswrapper[4778]: I0312 14:37:07.860728 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:37:07 crc kubenswrapper[4778]: I0312 14:37:07.912958 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:37:08 crc kubenswrapper[4778]: I0312 14:37:08.254076 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:37:08 crc kubenswrapper[4778]: E0312 14:37:08.254496 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:37:08 crc kubenswrapper[4778]: I0312 14:37:08.807876 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:37:08 crc kubenswrapper[4778]: I0312 14:37:08.866544 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xr4w"] Mar 12 14:37:10 crc kubenswrapper[4778]: I0312 14:37:10.749243 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6xr4w" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="registry-server" containerID="cri-o://0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53" gracePeriod=2 Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.362015 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.418743 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-catalog-content\") pod \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.419072 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-utilities\") pod \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.419305 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf5jk\" (UniqueName: \"kubernetes.io/projected/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-kube-api-access-sf5jk\") pod \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\" (UID: \"8cc6e361-05ad-401e-b5ab-2070ca8ec46c\") " Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.421599 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-utilities" (OuterVolumeSpecName: "utilities") pod "8cc6e361-05ad-401e-b5ab-2070ca8ec46c" (UID: "8cc6e361-05ad-401e-b5ab-2070ca8ec46c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.440495 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-kube-api-access-sf5jk" (OuterVolumeSpecName: "kube-api-access-sf5jk") pod "8cc6e361-05ad-401e-b5ab-2070ca8ec46c" (UID: "8cc6e361-05ad-401e-b5ab-2070ca8ec46c"). InnerVolumeSpecName "kube-api-access-sf5jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.522378 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.522425 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf5jk\" (UniqueName: \"kubernetes.io/projected/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-kube-api-access-sf5jk\") on node \"crc\" DevicePath \"\"" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.586955 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cc6e361-05ad-401e-b5ab-2070ca8ec46c" (UID: "8cc6e361-05ad-401e-b5ab-2070ca8ec46c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.625172 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc6e361-05ad-401e-b5ab-2070ca8ec46c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.762435 4778 generic.go:334] "Generic (PLEG): container finished" podID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerID="0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53" exitCode=0 Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.762478 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerDied","Data":"0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53"} Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.762505 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4w" event={"ID":"8cc6e361-05ad-401e-b5ab-2070ca8ec46c","Type":"ContainerDied","Data":"64e2681b11aea9607bf55a800b77d410d04225bde783d6a66b217bb9ae3cf27c"} Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.762530 4778 scope.go:117] "RemoveContainer" containerID="0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.762565 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4w" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.801918 4778 scope.go:117] "RemoveContainer" containerID="3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea" Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.812599 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xr4w"] Mar 12 14:37:11 crc kubenswrapper[4778]: I0312 14:37:11.820490 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6xr4w"] Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.264632 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" path="/var/lib/kubelet/pods/8cc6e361-05ad-401e-b5ab-2070ca8ec46c/volumes" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.437767 4778 scope.go:117] "RemoveContainer" containerID="71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.503041 4778 scope.go:117] "RemoveContainer" containerID="0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53" Mar 12 14:37:12 crc kubenswrapper[4778]: E0312 14:37:12.503495 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53\": container with ID starting with 0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53 not found: ID does not exist" containerID="0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.503548 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53"} err="failed to get container status \"0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53\": rpc error: code = NotFound desc = could not find container \"0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53\": container with ID starting with 0291482883d3905bcf0f1677176123126d7b804149fde1e81c7ae29d92b03d53 not found: ID does not exist" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.503580 4778 scope.go:117] "RemoveContainer" containerID="3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea" Mar 12 14:37:12 crc kubenswrapper[4778]: E0312 14:37:12.503879 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea\": container with ID starting with 3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea not found: ID does not exist" containerID="3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.503910 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea"} err="failed to get container status \"3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea\": rpc error: code = NotFound desc = could not find container \"3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea\": container with ID starting with 3f6e8c4ba226104d47c3abd66b373a0394b62e7f03fdaa0463457d44bb5412ea not found: ID does not exist" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.503929 4778 scope.go:117] "RemoveContainer" containerID="71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144" Mar 12 14:37:12 crc kubenswrapper[4778]: E0312 14:37:12.504150 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144\": container with ID starting with 71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144 not found: ID does not exist" containerID="71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144" Mar 12 14:37:12 crc kubenswrapper[4778]: I0312 14:37:12.504193 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144"} err="failed to get container status \"71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144\": rpc error: code = NotFound desc = could not find container \"71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144\": container with ID starting with 71033b90ec1c23b5bc0edc20bbd939347abc398d65871787dd9105b4cea19144 not found: ID does not exist" Mar 12 14:37:19 crc kubenswrapper[4778]: I0312 14:37:19.254766 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:37:19 crc kubenswrapper[4778]: E0312 14:37:19.256902 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:37:32 crc kubenswrapper[4778]: I0312 14:37:32.260319 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:37:32 crc kubenswrapper[4778]: E0312 14:37:32.261232 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:37:45 crc kubenswrapper[4778]: I0312 14:37:45.253756 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:37:45 crc kubenswrapper[4778]: E0312 14:37:45.254559 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.145304 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555438-86rh2"] Mar 12 14:38:00 crc kubenswrapper[4778]: E0312 14:38:00.146240 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="extract-utilities" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.146253 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="extract-utilities" Mar 12 14:38:00 crc kubenswrapper[4778]: E0312 14:38:00.146274 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="extract-content" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.146281 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="extract-content" Mar 12 14:38:00 crc kubenswrapper[4778]: E0312 14:38:00.146292 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="registry-server" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.146299 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="registry-server" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.146493 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc6e361-05ad-401e-b5ab-2070ca8ec46c" containerName="registry-server" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.147192 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.149861 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.149971 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.156903 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.158421 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555438-86rh2"] Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.254601 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:38:00 crc kubenswrapper[4778]: E0312 14:38:00.254854 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.281531 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95rwl\" (UniqueName: \"kubernetes.io/projected/bfdb06b1-1cad-4645-bf9c-1859648637ea-kube-api-access-95rwl\") pod \"auto-csr-approver-29555438-86rh2\" (UID: \"bfdb06b1-1cad-4645-bf9c-1859648637ea\") " pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.383976 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95rwl\" (UniqueName: \"kubernetes.io/projected/bfdb06b1-1cad-4645-bf9c-1859648637ea-kube-api-access-95rwl\") pod \"auto-csr-approver-29555438-86rh2\" (UID: \"bfdb06b1-1cad-4645-bf9c-1859648637ea\") " pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.405238 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95rwl\" (UniqueName: \"kubernetes.io/projected/bfdb06b1-1cad-4645-bf9c-1859648637ea-kube-api-access-95rwl\") pod \"auto-csr-approver-29555438-86rh2\" (UID: \"bfdb06b1-1cad-4645-bf9c-1859648637ea\") " pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.474919 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:00 crc kubenswrapper[4778]: I0312 14:38:00.912042 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555438-86rh2"] Mar 12 14:38:01 crc kubenswrapper[4778]: I0312 14:38:01.229196 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555438-86rh2" event={"ID":"bfdb06b1-1cad-4645-bf9c-1859648637ea","Type":"ContainerStarted","Data":"10c95db7de2a932f9562c52101bf7b1031b2e444269669dbb369190eab1fb105"} Mar 12 14:38:03 crc kubenswrapper[4778]: I0312 14:38:03.248559 4778 generic.go:334] "Generic (PLEG): container finished" podID="bfdb06b1-1cad-4645-bf9c-1859648637ea" containerID="a83e8d7acc08020b09631971d619c027f357278529f43520703cd7515d291f17" exitCode=0 Mar 12 14:38:03 crc kubenswrapper[4778]: I0312 14:38:03.248672 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555438-86rh2" event={"ID":"bfdb06b1-1cad-4645-bf9c-1859648637ea","Type":"ContainerDied","Data":"a83e8d7acc08020b09631971d619c027f357278529f43520703cd7515d291f17"} Mar 12 14:38:04 crc kubenswrapper[4778]: I0312 14:38:04.705687 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:04 crc kubenswrapper[4778]: I0312 14:38:04.867290 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95rwl\" (UniqueName: \"kubernetes.io/projected/bfdb06b1-1cad-4645-bf9c-1859648637ea-kube-api-access-95rwl\") pod \"bfdb06b1-1cad-4645-bf9c-1859648637ea\" (UID: \"bfdb06b1-1cad-4645-bf9c-1859648637ea\") " Mar 12 14:38:04 crc kubenswrapper[4778]: I0312 14:38:04.874053 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfdb06b1-1cad-4645-bf9c-1859648637ea-kube-api-access-95rwl" (OuterVolumeSpecName: "kube-api-access-95rwl") pod "bfdb06b1-1cad-4645-bf9c-1859648637ea" (UID: "bfdb06b1-1cad-4645-bf9c-1859648637ea"). InnerVolumeSpecName "kube-api-access-95rwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:38:04 crc kubenswrapper[4778]: I0312 14:38:04.970541 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95rwl\" (UniqueName: \"kubernetes.io/projected/bfdb06b1-1cad-4645-bf9c-1859648637ea-kube-api-access-95rwl\") on node \"crc\" DevicePath \"\"" Mar 12 14:38:05 crc kubenswrapper[4778]: I0312 14:38:05.267092 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555438-86rh2" event={"ID":"bfdb06b1-1cad-4645-bf9c-1859648637ea","Type":"ContainerDied","Data":"10c95db7de2a932f9562c52101bf7b1031b2e444269669dbb369190eab1fb105"} Mar 12 14:38:05 crc kubenswrapper[4778]: I0312 14:38:05.267146 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10c95db7de2a932f9562c52101bf7b1031b2e444269669dbb369190eab1fb105" Mar 12 14:38:05 crc kubenswrapper[4778]: I0312 14:38:05.267169 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555438-86rh2" Mar 12 14:38:05 crc kubenswrapper[4778]: I0312 14:38:05.787761 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555432-94dlz"] Mar 12 14:38:05 crc kubenswrapper[4778]: I0312 14:38:05.795770 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555432-94dlz"] Mar 12 14:38:06 crc kubenswrapper[4778]: I0312 14:38:06.268283 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b6391c3-533c-4b44-b1be-2a5c9752ba4b" path="/var/lib/kubelet/pods/4b6391c3-533c-4b44-b1be-2a5c9752ba4b/volumes" Mar 12 14:38:15 crc kubenswrapper[4778]: I0312 14:38:15.255212 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:38:15 crc kubenswrapper[4778]: E0312 14:38:15.256259 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:38:23 crc kubenswrapper[4778]: I0312 14:38:23.061501 4778 scope.go:117] "RemoveContainer" containerID="f68c8ed6b7c2e6259023580b179d97b5ef4d89ae76842473f005cc28f0933cea" Mar 12 14:38:30 crc kubenswrapper[4778]: I0312 14:38:30.254815 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:38:30 crc kubenswrapper[4778]: E0312 14:38:30.255789 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:38:43 crc kubenswrapper[4778]: I0312 14:38:43.254087 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:38:43 crc kubenswrapper[4778]: E0312 14:38:43.254769 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:38:54 crc kubenswrapper[4778]: I0312 14:38:54.254005 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:38:54 crc kubenswrapper[4778]: E0312 14:38:54.254722 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:39:09 crc kubenswrapper[4778]: I0312 14:39:09.254246 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:39:09 crc kubenswrapper[4778]: I0312 14:39:09.859438 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"7db21caa41aaa05f213157a5fdd43b948f849acb385674235e08738a115a03fb"} Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.146030 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555440-prm5s"] Mar 12 14:40:00 crc kubenswrapper[4778]: E0312 14:40:00.147699 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfdb06b1-1cad-4645-bf9c-1859648637ea" containerName="oc" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.147732 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfdb06b1-1cad-4645-bf9c-1859648637ea" containerName="oc" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.148273 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfdb06b1-1cad-4645-bf9c-1859648637ea" containerName="oc" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.149818 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.152025 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.152703 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.152899 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.162493 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555440-prm5s"] Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.285980 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/19be98d0-a8c0-4e30-926a-3ac799c6b576-kube-api-access-fxjbr\") pod \"auto-csr-approver-29555440-prm5s\" (UID: \"19be98d0-a8c0-4e30-926a-3ac799c6b576\") " pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.388415 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/19be98d0-a8c0-4e30-926a-3ac799c6b576-kube-api-access-fxjbr\") pod \"auto-csr-approver-29555440-prm5s\" (UID: \"19be98d0-a8c0-4e30-926a-3ac799c6b576\") " pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.409346 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/19be98d0-a8c0-4e30-926a-3ac799c6b576-kube-api-access-fxjbr\") pod \"auto-csr-approver-29555440-prm5s\" (UID: \"19be98d0-a8c0-4e30-926a-3ac799c6b576\") " pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.476936 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:00 crc kubenswrapper[4778]: I0312 14:40:00.914444 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555440-prm5s"] Mar 12 14:40:01 crc kubenswrapper[4778]: I0312 14:40:01.682222 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555440-prm5s" event={"ID":"19be98d0-a8c0-4e30-926a-3ac799c6b576","Type":"ContainerStarted","Data":"7aa5a18892e344dd85e64525777060db1300a26b3b462e75255e52fe17efa0d8"} Mar 12 14:40:02 crc kubenswrapper[4778]: I0312 14:40:02.692956 4778 generic.go:334] "Generic (PLEG): container finished" podID="19be98d0-a8c0-4e30-926a-3ac799c6b576" containerID="e11c1ed077d226d809927c1b3f1c2bde51c60c0a0093049ef2bd6c0eb78accb3" exitCode=0 Mar 12 14:40:02 crc kubenswrapper[4778]: I0312 14:40:02.693236 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555440-prm5s" event={"ID":"19be98d0-a8c0-4e30-926a-3ac799c6b576","Type":"ContainerDied","Data":"e11c1ed077d226d809927c1b3f1c2bde51c60c0a0093049ef2bd6c0eb78accb3"} Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.162754 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.284364 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/19be98d0-a8c0-4e30-926a-3ac799c6b576-kube-api-access-fxjbr\") pod \"19be98d0-a8c0-4e30-926a-3ac799c6b576\" (UID: \"19be98d0-a8c0-4e30-926a-3ac799c6b576\") " Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.290860 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19be98d0-a8c0-4e30-926a-3ac799c6b576-kube-api-access-fxjbr" (OuterVolumeSpecName: "kube-api-access-fxjbr") pod "19be98d0-a8c0-4e30-926a-3ac799c6b576" (UID: "19be98d0-a8c0-4e30-926a-3ac799c6b576"). InnerVolumeSpecName "kube-api-access-fxjbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.386864 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/19be98d0-a8c0-4e30-926a-3ac799c6b576-kube-api-access-fxjbr\") on node \"crc\" DevicePath \"\"" Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.712671 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555440-prm5s" event={"ID":"19be98d0-a8c0-4e30-926a-3ac799c6b576","Type":"ContainerDied","Data":"7aa5a18892e344dd85e64525777060db1300a26b3b462e75255e52fe17efa0d8"} Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.712710 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7aa5a18892e344dd85e64525777060db1300a26b3b462e75255e52fe17efa0d8" Mar 12 14:40:04 crc kubenswrapper[4778]: I0312 14:40:04.712724 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555440-prm5s" Mar 12 14:40:05 crc kubenswrapper[4778]: I0312 14:40:05.237874 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555434-hzddc"] Mar 12 14:40:05 crc kubenswrapper[4778]: I0312 14:40:05.246651 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555434-hzddc"] Mar 12 14:40:06 crc kubenswrapper[4778]: I0312 14:40:06.264959 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7855ae5-9f57-4d62-ab01-d16ae9f5a037" path="/var/lib/kubelet/pods/e7855ae5-9f57-4d62-ab01-d16ae9f5a037/volumes" Mar 12 14:40:23 crc kubenswrapper[4778]: I0312 14:40:23.175875 4778 scope.go:117] "RemoveContainer" containerID="ecf6cfdc210df01866b5bda8e874db3a9407a84531517ea05fb802b9d57bcdb0" Mar 12 14:41:28 crc kubenswrapper[4778]: I0312 14:41:28.557636 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:41:28 crc kubenswrapper[4778]: I0312 14:41:28.558286 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.382918 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ghxk4"] Mar 12 14:41:41 crc kubenswrapper[4778]: E0312 14:41:41.384119 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19be98d0-a8c0-4e30-926a-3ac799c6b576" containerName="oc" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.384243 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="19be98d0-a8c0-4e30-926a-3ac799c6b576" containerName="oc" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.384510 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="19be98d0-a8c0-4e30-926a-3ac799c6b576" containerName="oc" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.386228 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.400976 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghxk4"] Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.521978 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-utilities\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.522066 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdrj9\" (UniqueName: \"kubernetes.io/projected/08484b9b-c7e4-4119-b306-ba766bfdab7c-kube-api-access-pdrj9\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.522173 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-catalog-content\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.624824 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-catalog-content\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.624927 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-utilities\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.624990 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdrj9\" (UniqueName: \"kubernetes.io/projected/08484b9b-c7e4-4119-b306-ba766bfdab7c-kube-api-access-pdrj9\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.625525 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-catalog-content\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.625568 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-utilities\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.652962 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdrj9\" (UniqueName: \"kubernetes.io/projected/08484b9b-c7e4-4119-b306-ba766bfdab7c-kube-api-access-pdrj9\") pod \"redhat-marketplace-ghxk4\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:41 crc kubenswrapper[4778]: I0312 14:41:41.715818 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:42 crc kubenswrapper[4778]: I0312 14:41:42.247584 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghxk4"] Mar 12 14:41:43 crc kubenswrapper[4778]: I0312 14:41:43.212291 4778 generic.go:334] "Generic (PLEG): container finished" podID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerID="fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0" exitCode=0 Mar 12 14:41:43 crc kubenswrapper[4778]: I0312 14:41:43.212395 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerDied","Data":"fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0"} Mar 12 14:41:43 crc kubenswrapper[4778]: I0312 14:41:43.212664 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerStarted","Data":"257844a9141576a3172cff81ed4e5381828b5b4bc86357e1a4f01ef234ab7c96"} Mar 12 14:41:43 crc kubenswrapper[4778]: I0312 14:41:43.215368 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:41:44 crc kubenswrapper[4778]: I0312 14:41:44.224938 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerStarted","Data":"9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff"} Mar 12 14:41:45 crc kubenswrapper[4778]: E0312 14:41:45.112268 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08484b9b_c7e4_4119_b306_ba766bfdab7c.slice/crio-conmon-9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08484b9b_c7e4_4119_b306_ba766bfdab7c.slice/crio-9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff.scope\": RecentStats: unable to find data in memory cache]" Mar 12 14:41:45 crc kubenswrapper[4778]: I0312 14:41:45.235099 4778 generic.go:334] "Generic (PLEG): container finished" podID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerID="9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff" exitCode=0 Mar 12 14:41:45 crc kubenswrapper[4778]: I0312 14:41:45.235156 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerDied","Data":"9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff"} Mar 12 14:41:46 crc kubenswrapper[4778]: I0312 14:41:46.243978 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerStarted","Data":"63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763"} Mar 12 14:41:46 crc kubenswrapper[4778]: I0312 14:41:46.259625 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ghxk4" podStartSLOduration=2.595066854 podStartE2EDuration="5.259608188s" podCreationTimestamp="2026-03-12 14:41:41 +0000 UTC" firstStartedPulling="2026-03-12 14:41:43.215071249 +0000 UTC m=+5521.663766645" lastFinishedPulling="2026-03-12 14:41:45.879612563 +0000 UTC m=+5524.328307979" observedRunningTime="2026-03-12 14:41:46.25895518 +0000 UTC m=+5524.707650576" watchObservedRunningTime="2026-03-12 14:41:46.259608188 +0000 UTC m=+5524.708303584" Mar 12 14:41:51 crc kubenswrapper[4778]: I0312 14:41:51.716488 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:51 crc kubenswrapper[4778]: I0312 14:41:51.718381 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:51 crc kubenswrapper[4778]: I0312 14:41:51.773746 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:52 crc kubenswrapper[4778]: I0312 14:41:52.356705 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:52 crc kubenswrapper[4778]: I0312 14:41:52.401252 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghxk4"] Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.317413 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ghxk4" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="registry-server" containerID="cri-o://63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763" gracePeriod=2 Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.753245 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.787609 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-utilities\") pod \"08484b9b-c7e4-4119-b306-ba766bfdab7c\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.787697 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdrj9\" (UniqueName: \"kubernetes.io/projected/08484b9b-c7e4-4119-b306-ba766bfdab7c-kube-api-access-pdrj9\") pod \"08484b9b-c7e4-4119-b306-ba766bfdab7c\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.787766 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-catalog-content\") pod \"08484b9b-c7e4-4119-b306-ba766bfdab7c\" (UID: \"08484b9b-c7e4-4119-b306-ba766bfdab7c\") " Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.794116 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-utilities" (OuterVolumeSpecName: "utilities") pod "08484b9b-c7e4-4119-b306-ba766bfdab7c" (UID: "08484b9b-c7e4-4119-b306-ba766bfdab7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.805485 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08484b9b-c7e4-4119-b306-ba766bfdab7c-kube-api-access-pdrj9" (OuterVolumeSpecName: "kube-api-access-pdrj9") pod "08484b9b-c7e4-4119-b306-ba766bfdab7c" (UID: "08484b9b-c7e4-4119-b306-ba766bfdab7c"). InnerVolumeSpecName "kube-api-access-pdrj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.848586 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08484b9b-c7e4-4119-b306-ba766bfdab7c" (UID: "08484b9b-c7e4-4119-b306-ba766bfdab7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.890767 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.890802 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdrj9\" (UniqueName: \"kubernetes.io/projected/08484b9b-c7e4-4119-b306-ba766bfdab7c-kube-api-access-pdrj9\") on node \"crc\" DevicePath \"\"" Mar 12 14:41:54 crc kubenswrapper[4778]: I0312 14:41:54.890816 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08484b9b-c7e4-4119-b306-ba766bfdab7c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.328472 4778 generic.go:334] "Generic (PLEG): container finished" podID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerID="63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763" exitCode=0 Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.328516 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerDied","Data":"63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763"} Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.328540 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghxk4" event={"ID":"08484b9b-c7e4-4119-b306-ba766bfdab7c","Type":"ContainerDied","Data":"257844a9141576a3172cff81ed4e5381828b5b4bc86357e1a4f01ef234ab7c96"} Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.328554 4778 scope.go:117] "RemoveContainer" containerID="63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.328670 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghxk4" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.350026 4778 scope.go:117] "RemoveContainer" containerID="9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.365577 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghxk4"] Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.374490 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghxk4"] Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.399821 4778 scope.go:117] "RemoveContainer" containerID="fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.426544 4778 scope.go:117] "RemoveContainer" containerID="63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763" Mar 12 14:41:55 crc kubenswrapper[4778]: E0312 14:41:55.426764 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763\": container with ID starting with 63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763 not found: ID does not exist" containerID="63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.426795 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763"} err="failed to get container status \"63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763\": rpc error: code = NotFound desc = could not find container \"63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763\": container with ID starting with 63ec0182dbfdb3076f9264015703df928906cfd80a27db11b18bb38d83978763 not found: ID does not exist" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.426817 4778 scope.go:117] "RemoveContainer" containerID="9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff" Mar 12 14:41:55 crc kubenswrapper[4778]: E0312 14:41:55.426961 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff\": container with ID starting with 9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff not found: ID does not exist" containerID="9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.426982 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff"} err="failed to get container status \"9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff\": rpc error: code = NotFound desc = could not find container \"9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff\": container with ID starting with 9cc32ffde6bb29a0fb51938eff56d0c36bc993785d3b7a77680e754bca9602ff not found: ID does not exist" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.426992 4778 scope.go:117] "RemoveContainer" containerID="fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0" Mar 12 14:41:55 crc kubenswrapper[4778]: E0312 14:41:55.427205 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0\": container with ID starting with fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0 not found: ID does not exist" containerID="fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0" Mar 12 14:41:55 crc kubenswrapper[4778]: I0312 14:41:55.427222 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0"} err="failed to get container status \"fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0\": rpc error: code = NotFound desc = could not find container \"fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0\": container with ID starting with fbeb0cd20ee047ff32cbe65ff8cd88b9cd6a713b1ac825933fe8d93c01221de0 not found: ID does not exist" Mar 12 14:41:56 crc kubenswrapper[4778]: I0312 14:41:56.264121 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" path="/var/lib/kubelet/pods/08484b9b-c7e4-4119-b306-ba766bfdab7c/volumes" Mar 12 14:41:58 crc kubenswrapper[4778]: I0312 14:41:58.558123 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:41:58 crc kubenswrapper[4778]: I0312 14:41:58.558467 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.146166 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555442-c58f5"] Mar 12 14:42:00 crc kubenswrapper[4778]: E0312 14:42:00.147031 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="extract-utilities" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.147050 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="extract-utilities" Mar 12 14:42:00 crc kubenswrapper[4778]: E0312 14:42:00.147087 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="registry-server" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.147095 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="registry-server" Mar 12 14:42:00 crc kubenswrapper[4778]: E0312 14:42:00.147131 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="extract-content" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.147138 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="extract-content" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.147383 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="08484b9b-c7e4-4119-b306-ba766bfdab7c" containerName="registry-server" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.148199 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.153475 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.153686 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.153792 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.171543 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555442-c58f5"] Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.192853 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp8b6\" (UniqueName: \"kubernetes.io/projected/86de30c2-8699-4966-8c1b-da67cdacae42-kube-api-access-dp8b6\") pod \"auto-csr-approver-29555442-c58f5\" (UID: \"86de30c2-8699-4966-8c1b-da67cdacae42\") " pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.295046 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp8b6\" (UniqueName: \"kubernetes.io/projected/86de30c2-8699-4966-8c1b-da67cdacae42-kube-api-access-dp8b6\") pod \"auto-csr-approver-29555442-c58f5\" (UID: \"86de30c2-8699-4966-8c1b-da67cdacae42\") " pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.317395 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp8b6\" (UniqueName: \"kubernetes.io/projected/86de30c2-8699-4966-8c1b-da67cdacae42-kube-api-access-dp8b6\") pod \"auto-csr-approver-29555442-c58f5\" (UID: \"86de30c2-8699-4966-8c1b-da67cdacae42\") " pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.469410 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:00 crc kubenswrapper[4778]: I0312 14:42:00.930593 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555442-c58f5"] Mar 12 14:42:01 crc kubenswrapper[4778]: I0312 14:42:01.375160 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555442-c58f5" event={"ID":"86de30c2-8699-4966-8c1b-da67cdacae42","Type":"ContainerStarted","Data":"e9ca4027e65d1eece4dde85d28cad4df3531606fa0be1367675a67dabdc4867a"} Mar 12 14:42:03 crc kubenswrapper[4778]: I0312 14:42:03.395896 4778 generic.go:334] "Generic (PLEG): container finished" podID="86de30c2-8699-4966-8c1b-da67cdacae42" containerID="629d070304f0ca91f60ef09f2871ae160406fbd39c685129feb137e8e63e7888" exitCode=0 Mar 12 14:42:03 crc kubenswrapper[4778]: I0312 14:42:03.395934 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555442-c58f5" event={"ID":"86de30c2-8699-4966-8c1b-da67cdacae42","Type":"ContainerDied","Data":"629d070304f0ca91f60ef09f2871ae160406fbd39c685129feb137e8e63e7888"} Mar 12 14:42:04 crc kubenswrapper[4778]: I0312 14:42:04.735378 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:04 crc kubenswrapper[4778]: I0312 14:42:04.785241 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp8b6\" (UniqueName: \"kubernetes.io/projected/86de30c2-8699-4966-8c1b-da67cdacae42-kube-api-access-dp8b6\") pod \"86de30c2-8699-4966-8c1b-da67cdacae42\" (UID: \"86de30c2-8699-4966-8c1b-da67cdacae42\") " Mar 12 14:42:04 crc kubenswrapper[4778]: I0312 14:42:04.795117 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86de30c2-8699-4966-8c1b-da67cdacae42-kube-api-access-dp8b6" (OuterVolumeSpecName: "kube-api-access-dp8b6") pod "86de30c2-8699-4966-8c1b-da67cdacae42" (UID: "86de30c2-8699-4966-8c1b-da67cdacae42"). InnerVolumeSpecName "kube-api-access-dp8b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:42:04 crc kubenswrapper[4778]: I0312 14:42:04.889670 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp8b6\" (UniqueName: \"kubernetes.io/projected/86de30c2-8699-4966-8c1b-da67cdacae42-kube-api-access-dp8b6\") on node \"crc\" DevicePath \"\"" Mar 12 14:42:05 crc kubenswrapper[4778]: I0312 14:42:05.414472 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555442-c58f5" event={"ID":"86de30c2-8699-4966-8c1b-da67cdacae42","Type":"ContainerDied","Data":"e9ca4027e65d1eece4dde85d28cad4df3531606fa0be1367675a67dabdc4867a"} Mar 12 14:42:05 crc kubenswrapper[4778]: I0312 14:42:05.414511 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9ca4027e65d1eece4dde85d28cad4df3531606fa0be1367675a67dabdc4867a" Mar 12 14:42:05 crc kubenswrapper[4778]: I0312 14:42:05.414565 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555442-c58f5" Mar 12 14:42:05 crc kubenswrapper[4778]: E0312 14:42:05.601954 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86de30c2_8699_4966_8c1b_da67cdacae42.slice/crio-e9ca4027e65d1eece4dde85d28cad4df3531606fa0be1367675a67dabdc4867a\": RecentStats: unable to find data in memory cache]" Mar 12 14:42:05 crc kubenswrapper[4778]: I0312 14:42:05.859197 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555436-9bs4m"] Mar 12 14:42:05 crc kubenswrapper[4778]: I0312 14:42:05.878897 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555436-9bs4m"] Mar 12 14:42:06 crc kubenswrapper[4778]: I0312 14:42:06.266253 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d1f962-b71e-4473-b387-137a395e1a39" path="/var/lib/kubelet/pods/45d1f962-b71e-4473-b387-137a395e1a39/volumes" Mar 12 14:42:23 crc kubenswrapper[4778]: I0312 14:42:23.270420 4778 scope.go:117] "RemoveContainer" containerID="b4d039fad9b993f652c5f6f0f661f085d4f93b47467c47a0fe13959b9f367b5d" Mar 12 14:42:28 crc kubenswrapper[4778]: I0312 14:42:28.557810 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:42:28 crc kubenswrapper[4778]: I0312 14:42:28.559510 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:42:28 crc kubenswrapper[4778]: I0312 14:42:28.559684 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:42:28 crc kubenswrapper[4778]: I0312 14:42:28.560604 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7db21caa41aaa05f213157a5fdd43b948f849acb385674235e08738a115a03fb"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:42:28 crc kubenswrapper[4778]: I0312 14:42:28.560751 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://7db21caa41aaa05f213157a5fdd43b948f849acb385674235e08738a115a03fb" gracePeriod=600 Mar 12 14:42:29 crc kubenswrapper[4778]: I0312 14:42:29.047329 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="7db21caa41aaa05f213157a5fdd43b948f849acb385674235e08738a115a03fb" exitCode=0 Mar 12 14:42:29 crc kubenswrapper[4778]: I0312 14:42:29.047826 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"7db21caa41aaa05f213157a5fdd43b948f849acb385674235e08738a115a03fb"} Mar 12 14:42:29 crc kubenswrapper[4778]: I0312 14:42:29.047855 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4"} Mar 12 14:42:29 crc kubenswrapper[4778]: I0312 14:42:29.047872 4778 scope.go:117] "RemoveContainer" containerID="bbb7cd318ed9aaf8c81b44eaf9e441283227b353d96ce94d2989c6c892e1351c" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.169868 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mqtgd"] Mar 12 14:42:44 crc kubenswrapper[4778]: E0312 14:42:44.171392 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86de30c2-8699-4966-8c1b-da67cdacae42" containerName="oc" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.171414 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="86de30c2-8699-4966-8c1b-da67cdacae42" containerName="oc" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.171750 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="86de30c2-8699-4966-8c1b-da67cdacae42" containerName="oc" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.173520 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.185736 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqtgd"] Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.221927 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-utilities\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.221998 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-catalog-content\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.222288 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86g7c\" (UniqueName: \"kubernetes.io/projected/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-kube-api-access-86g7c\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.324841 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86g7c\" (UniqueName: \"kubernetes.io/projected/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-kube-api-access-86g7c\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.325048 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-utilities\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.325077 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-catalog-content\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.325996 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-catalog-content\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.326074 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-utilities\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.356068 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86g7c\" (UniqueName: \"kubernetes.io/projected/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-kube-api-access-86g7c\") pod \"redhat-operators-mqtgd\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:44 crc kubenswrapper[4778]: I0312 14:42:44.493691 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:45 crc kubenswrapper[4778]: I0312 14:42:44.998782 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqtgd"] Mar 12 14:42:45 crc kubenswrapper[4778]: I0312 14:42:45.187242 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerStarted","Data":"f6a4fb35fc6694b18950498d5d09e6c3816748fe0f6054bbd9a78d05234e573b"} Mar 12 14:42:46 crc kubenswrapper[4778]: I0312 14:42:46.196727 4778 generic.go:334] "Generic (PLEG): container finished" podID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerID="58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad" exitCode=0 Mar 12 14:42:46 crc kubenswrapper[4778]: I0312 14:42:46.196773 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerDied","Data":"58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad"} Mar 12 14:42:47 crc kubenswrapper[4778]: I0312 14:42:47.210090 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerStarted","Data":"cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623"} Mar 12 14:42:52 crc kubenswrapper[4778]: I0312 14:42:52.265452 4778 generic.go:334] "Generic (PLEG): container finished" podID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerID="cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623" exitCode=0 Mar 12 14:42:52 crc kubenswrapper[4778]: I0312 14:42:52.267610 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerDied","Data":"cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623"} Mar 12 14:42:53 crc kubenswrapper[4778]: I0312 14:42:53.280573 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerStarted","Data":"dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631"} Mar 12 14:42:53 crc kubenswrapper[4778]: I0312 14:42:53.321827 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mqtgd" podStartSLOduration=2.763196979 podStartE2EDuration="9.321802669s" podCreationTimestamp="2026-03-12 14:42:44 +0000 UTC" firstStartedPulling="2026-03-12 14:42:46.19911342 +0000 UTC m=+5584.647808816" lastFinishedPulling="2026-03-12 14:42:52.75771907 +0000 UTC m=+5591.206414506" observedRunningTime="2026-03-12 14:42:53.307254396 +0000 UTC m=+5591.755949842" watchObservedRunningTime="2026-03-12 14:42:53.321802669 +0000 UTC m=+5591.770498075" Mar 12 14:42:54 crc kubenswrapper[4778]: I0312 14:42:54.494234 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:54 crc kubenswrapper[4778]: I0312 14:42:54.494561 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:42:55 crc kubenswrapper[4778]: I0312 14:42:55.553774 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mqtgd" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="registry-server" probeResult="failure" output=< Mar 12 14:42:55 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:42:55 crc kubenswrapper[4778]: > Mar 12 14:43:04 crc kubenswrapper[4778]: I0312 14:43:04.543991 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:43:04 crc kubenswrapper[4778]: I0312 14:43:04.591705 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:43:04 crc kubenswrapper[4778]: I0312 14:43:04.784222 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqtgd"] Mar 12 14:43:06 crc kubenswrapper[4778]: I0312 14:43:06.395918 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mqtgd" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="registry-server" containerID="cri-o://dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631" gracePeriod=2 Mar 12 14:43:06 crc kubenswrapper[4778]: I0312 14:43:06.942101 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.010239 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-catalog-content\") pod \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.010531 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-utilities\") pod \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.010654 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86g7c\" (UniqueName: \"kubernetes.io/projected/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-kube-api-access-86g7c\") pod \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\" (UID: \"bcd4d515-136b-47e8-92ba-9e79ed98e8ec\") " Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.012351 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-utilities" (OuterVolumeSpecName: "utilities") pod "bcd4d515-136b-47e8-92ba-9e79ed98e8ec" (UID: "bcd4d515-136b-47e8-92ba-9e79ed98e8ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.021327 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-kube-api-access-86g7c" (OuterVolumeSpecName: "kube-api-access-86g7c") pod "bcd4d515-136b-47e8-92ba-9e79ed98e8ec" (UID: "bcd4d515-136b-47e8-92ba-9e79ed98e8ec"). InnerVolumeSpecName "kube-api-access-86g7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.113018 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.113055 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86g7c\" (UniqueName: \"kubernetes.io/projected/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-kube-api-access-86g7c\") on node \"crc\" DevicePath \"\"" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.172963 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcd4d515-136b-47e8-92ba-9e79ed98e8ec" (UID: "bcd4d515-136b-47e8-92ba-9e79ed98e8ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.214870 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd4d515-136b-47e8-92ba-9e79ed98e8ec-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.405559 4778 generic.go:334] "Generic (PLEG): container finished" podID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerID="dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631" exitCode=0 Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.405619 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerDied","Data":"dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631"} Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.405645 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqtgd" event={"ID":"bcd4d515-136b-47e8-92ba-9e79ed98e8ec","Type":"ContainerDied","Data":"f6a4fb35fc6694b18950498d5d09e6c3816748fe0f6054bbd9a78d05234e573b"} Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.405662 4778 scope.go:117] "RemoveContainer" containerID="dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.405624 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqtgd" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.433915 4778 scope.go:117] "RemoveContainer" containerID="cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.466777 4778 scope.go:117] "RemoveContainer" containerID="58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.468960 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqtgd"] Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.480819 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mqtgd"] Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.520792 4778 scope.go:117] "RemoveContainer" containerID="dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631" Mar 12 14:43:07 crc kubenswrapper[4778]: E0312 14:43:07.521277 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631\": container with ID starting with dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631 not found: ID does not exist" containerID="dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.521405 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631"} err="failed to get container status \"dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631\": rpc error: code = NotFound desc = could not find container \"dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631\": container with ID starting with dacb248b27337d412ad2a8d93ae1a608ba1411ee25700ca3e1e7a3f5ff29d631 not found: ID does not exist" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.521529 4778 scope.go:117] "RemoveContainer" containerID="cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623" Mar 12 14:43:07 crc kubenswrapper[4778]: E0312 14:43:07.521953 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623\": container with ID starting with cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623 not found: ID does not exist" containerID="cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.521988 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623"} err="failed to get container status \"cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623\": rpc error: code = NotFound desc = could not find container \"cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623\": container with ID starting with cbd788e3627ea03805a19638b323850395483e01b39b06317c94c025d0b17623 not found: ID does not exist" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.522011 4778 scope.go:117] "RemoveContainer" containerID="58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad" Mar 12 14:43:07 crc kubenswrapper[4778]: E0312 14:43:07.522236 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad\": container with ID starting with 58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad not found: ID does not exist" containerID="58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad" Mar 12 14:43:07 crc kubenswrapper[4778]: I0312 14:43:07.522267 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad"} err="failed to get container status \"58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad\": rpc error: code = NotFound desc = could not find container \"58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad\": container with ID starting with 58941a5bd25d7707244796e0457ad9e6c2eecea41a390bebb99271f9e229c8ad not found: ID does not exist" Mar 12 14:43:08 crc kubenswrapper[4778]: I0312 14:43:08.268280 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" path="/var/lib/kubelet/pods/bcd4d515-136b-47e8-92ba-9e79ed98e8ec/volumes" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.152250 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555444-qlvql"] Mar 12 14:44:00 crc kubenswrapper[4778]: E0312 14:44:00.153584 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="extract-content" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.153603 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="extract-content" Mar 12 14:44:00 crc kubenswrapper[4778]: E0312 14:44:00.153627 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="extract-utilities" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.153636 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="extract-utilities" Mar 12 14:44:00 crc kubenswrapper[4778]: E0312 14:44:00.153667 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="registry-server" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.153678 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="registry-server" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.153948 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd4d515-136b-47e8-92ba-9e79ed98e8ec" containerName="registry-server" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.154822 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.158130 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.158607 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.160463 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.161969 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555444-qlvql"] Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.287475 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42fdw\" (UniqueName: \"kubernetes.io/projected/0da8caed-335b-4730-a69b-b724585893f0-kube-api-access-42fdw\") pod \"auto-csr-approver-29555444-qlvql\" (UID: \"0da8caed-335b-4730-a69b-b724585893f0\") " pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.389896 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42fdw\" (UniqueName: \"kubernetes.io/projected/0da8caed-335b-4730-a69b-b724585893f0-kube-api-access-42fdw\") pod \"auto-csr-approver-29555444-qlvql\" (UID: \"0da8caed-335b-4730-a69b-b724585893f0\") " pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.412072 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42fdw\" (UniqueName: \"kubernetes.io/projected/0da8caed-335b-4730-a69b-b724585893f0-kube-api-access-42fdw\") pod \"auto-csr-approver-29555444-qlvql\" (UID: \"0da8caed-335b-4730-a69b-b724585893f0\") " pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.484672 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:00 crc kubenswrapper[4778]: I0312 14:44:00.935574 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555444-qlvql"] Mar 12 14:44:01 crc kubenswrapper[4778]: I0312 14:44:01.890270 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555444-qlvql" event={"ID":"0da8caed-335b-4730-a69b-b724585893f0","Type":"ContainerStarted","Data":"b6dbcb7a022a068a1c8c79cecb2f2c3a82311a6d05a961f203bb6f31cc1279bb"} Mar 12 14:44:02 crc kubenswrapper[4778]: I0312 14:44:02.899235 4778 generic.go:334] "Generic (PLEG): container finished" podID="0da8caed-335b-4730-a69b-b724585893f0" containerID="1c045d93602ba35174db559893247fb3da6916ea0a208015b77cd0f47781b091" exitCode=0 Mar 12 14:44:02 crc kubenswrapper[4778]: I0312 14:44:02.899322 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555444-qlvql" event={"ID":"0da8caed-335b-4730-a69b-b724585893f0","Type":"ContainerDied","Data":"1c045d93602ba35174db559893247fb3da6916ea0a208015b77cd0f47781b091"} Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.296437 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.370878 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42fdw\" (UniqueName: \"kubernetes.io/projected/0da8caed-335b-4730-a69b-b724585893f0-kube-api-access-42fdw\") pod \"0da8caed-335b-4730-a69b-b724585893f0\" (UID: \"0da8caed-335b-4730-a69b-b724585893f0\") " Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.377695 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da8caed-335b-4730-a69b-b724585893f0-kube-api-access-42fdw" (OuterVolumeSpecName: "kube-api-access-42fdw") pod "0da8caed-335b-4730-a69b-b724585893f0" (UID: "0da8caed-335b-4730-a69b-b724585893f0"). InnerVolumeSpecName "kube-api-access-42fdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.473500 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42fdw\" (UniqueName: \"kubernetes.io/projected/0da8caed-335b-4730-a69b-b724585893f0-kube-api-access-42fdw\") on node \"crc\" DevicePath \"\"" Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.919257 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555444-qlvql" event={"ID":"0da8caed-335b-4730-a69b-b724585893f0","Type":"ContainerDied","Data":"b6dbcb7a022a068a1c8c79cecb2f2c3a82311a6d05a961f203bb6f31cc1279bb"} Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.919297 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6dbcb7a022a068a1c8c79cecb2f2c3a82311a6d05a961f203bb6f31cc1279bb" Mar 12 14:44:04 crc kubenswrapper[4778]: I0312 14:44:04.919581 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555444-qlvql" Mar 12 14:44:05 crc kubenswrapper[4778]: I0312 14:44:05.394523 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555438-86rh2"] Mar 12 14:44:05 crc kubenswrapper[4778]: I0312 14:44:05.406321 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555438-86rh2"] Mar 12 14:44:06 crc kubenswrapper[4778]: I0312 14:44:06.269330 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfdb06b1-1cad-4645-bf9c-1859648637ea" path="/var/lib/kubelet/pods/bfdb06b1-1cad-4645-bf9c-1859648637ea/volumes" Mar 12 14:44:23 crc kubenswrapper[4778]: I0312 14:44:23.414120 4778 scope.go:117] "RemoveContainer" containerID="a83e8d7acc08020b09631971d619c027f357278529f43520703cd7515d291f17" Mar 12 14:44:28 crc kubenswrapper[4778]: I0312 14:44:28.557579 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:44:28 crc kubenswrapper[4778]: I0312 14:44:28.558053 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:44:58 crc kubenswrapper[4778]: I0312 14:44:58.694631 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:44:58 crc kubenswrapper[4778]: I0312 14:44:58.695230 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.173317 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc"] Mar 12 14:45:00 crc kubenswrapper[4778]: E0312 14:45:00.175317 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da8caed-335b-4730-a69b-b724585893f0" containerName="oc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.175430 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da8caed-335b-4730-a69b-b724585893f0" containerName="oc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.175794 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0da8caed-335b-4730-a69b-b724585893f0" containerName="oc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.176751 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.179895 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.180301 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.186298 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc"] Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.325740 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a937d3c8-3521-400c-9703-ff806cd36e1f-secret-volume\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.325867 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a937d3c8-3521-400c-9703-ff806cd36e1f-config-volume\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.325982 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtf5n\" (UniqueName: \"kubernetes.io/projected/a937d3c8-3521-400c-9703-ff806cd36e1f-kube-api-access-vtf5n\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.427437 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtf5n\" (UniqueName: \"kubernetes.io/projected/a937d3c8-3521-400c-9703-ff806cd36e1f-kube-api-access-vtf5n\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.427657 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a937d3c8-3521-400c-9703-ff806cd36e1f-secret-volume\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.427699 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a937d3c8-3521-400c-9703-ff806cd36e1f-config-volume\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.428825 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a937d3c8-3521-400c-9703-ff806cd36e1f-config-volume\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.436491 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a937d3c8-3521-400c-9703-ff806cd36e1f-secret-volume\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.444400 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtf5n\" (UniqueName: \"kubernetes.io/projected/a937d3c8-3521-400c-9703-ff806cd36e1f-kube-api-access-vtf5n\") pod \"collect-profiles-29555445-h75tc\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.509076 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:00 crc kubenswrapper[4778]: I0312 14:45:00.990570 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc"] Mar 12 14:45:01 crc kubenswrapper[4778]: I0312 14:45:01.436857 4778 generic.go:334] "Generic (PLEG): container finished" podID="a937d3c8-3521-400c-9703-ff806cd36e1f" containerID="04cbbe7f2d03e6a5c907c38fa94f13fc54432108a3a87359ee7c5a5ffeb55d0e" exitCode=0 Mar 12 14:45:01 crc kubenswrapper[4778]: I0312 14:45:01.436922 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" event={"ID":"a937d3c8-3521-400c-9703-ff806cd36e1f","Type":"ContainerDied","Data":"04cbbe7f2d03e6a5c907c38fa94f13fc54432108a3a87359ee7c5a5ffeb55d0e"} Mar 12 14:45:01 crc kubenswrapper[4778]: I0312 14:45:01.436951 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" event={"ID":"a937d3c8-3521-400c-9703-ff806cd36e1f","Type":"ContainerStarted","Data":"a43137fc72539a88663622748f24aaf1646eda8d5c73c48ba4c842b379097cc1"} Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.833433 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.976086 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a937d3c8-3521-400c-9703-ff806cd36e1f-secret-volume\") pod \"a937d3c8-3521-400c-9703-ff806cd36e1f\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.976648 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a937d3c8-3521-400c-9703-ff806cd36e1f-config-volume\") pod \"a937d3c8-3521-400c-9703-ff806cd36e1f\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.977022 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtf5n\" (UniqueName: \"kubernetes.io/projected/a937d3c8-3521-400c-9703-ff806cd36e1f-kube-api-access-vtf5n\") pod \"a937d3c8-3521-400c-9703-ff806cd36e1f\" (UID: \"a937d3c8-3521-400c-9703-ff806cd36e1f\") " Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.977256 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a937d3c8-3521-400c-9703-ff806cd36e1f-config-volume" (OuterVolumeSpecName: "config-volume") pod "a937d3c8-3521-400c-9703-ff806cd36e1f" (UID: "a937d3c8-3521-400c-9703-ff806cd36e1f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.978064 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a937d3c8-3521-400c-9703-ff806cd36e1f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.982121 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a937d3c8-3521-400c-9703-ff806cd36e1f-kube-api-access-vtf5n" (OuterVolumeSpecName: "kube-api-access-vtf5n") pod "a937d3c8-3521-400c-9703-ff806cd36e1f" (UID: "a937d3c8-3521-400c-9703-ff806cd36e1f"). InnerVolumeSpecName "kube-api-access-vtf5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:45:02 crc kubenswrapper[4778]: I0312 14:45:02.982564 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a937d3c8-3521-400c-9703-ff806cd36e1f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a937d3c8-3521-400c-9703-ff806cd36e1f" (UID: "a937d3c8-3521-400c-9703-ff806cd36e1f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.080379 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a937d3c8-3521-400c-9703-ff806cd36e1f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.080420 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtf5n\" (UniqueName: \"kubernetes.io/projected/a937d3c8-3521-400c-9703-ff806cd36e1f-kube-api-access-vtf5n\") on node \"crc\" DevicePath \"\"" Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.462638 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" event={"ID":"a937d3c8-3521-400c-9703-ff806cd36e1f","Type":"ContainerDied","Data":"a43137fc72539a88663622748f24aaf1646eda8d5c73c48ba4c842b379097cc1"} Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.462695 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a43137fc72539a88663622748f24aaf1646eda8d5c73c48ba4c842b379097cc1" Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.462696 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555445-h75tc" Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.923613 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8"] Mar 12 14:45:03 crc kubenswrapper[4778]: I0312 14:45:03.934424 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555400-lrxd8"] Mar 12 14:45:04 crc kubenswrapper[4778]: I0312 14:45:04.266582 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d85560c-89e4-4723-beb0-aeda87d0791a" path="/var/lib/kubelet/pods/5d85560c-89e4-4723-beb0-aeda87d0791a/volumes" Mar 12 14:45:23 crc kubenswrapper[4778]: I0312 14:45:23.550111 4778 scope.go:117] "RemoveContainer" containerID="96aa4949ff208afd2c193ba8303ee15ee08731bdac5eecb0faaa4ff029a2c93a" Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.557846 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.558551 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.558637 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.559850 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.559972 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" gracePeriod=600 Mar 12 14:45:28 crc kubenswrapper[4778]: E0312 14:45:28.689636 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.732983 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" exitCode=0 Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.733028 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4"} Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.733094 4778 scope.go:117] "RemoveContainer" containerID="7db21caa41aaa05f213157a5fdd43b948f849acb385674235e08738a115a03fb" Mar 12 14:45:28 crc kubenswrapper[4778]: I0312 14:45:28.733622 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:45:28 crc kubenswrapper[4778]: E0312 14:45:28.733907 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:45:43 crc kubenswrapper[4778]: I0312 14:45:43.254564 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:45:43 crc kubenswrapper[4778]: E0312 14:45:43.255823 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:45:54 crc kubenswrapper[4778]: I0312 14:45:54.253784 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:45:54 crc kubenswrapper[4778]: E0312 14:45:54.254615 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.151145 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555446-8n794"] Mar 12 14:46:00 crc kubenswrapper[4778]: E0312 14:46:00.152126 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a937d3c8-3521-400c-9703-ff806cd36e1f" containerName="collect-profiles" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.152141 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="a937d3c8-3521-400c-9703-ff806cd36e1f" containerName="collect-profiles" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.152364 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="a937d3c8-3521-400c-9703-ff806cd36e1f" containerName="collect-profiles" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.152943 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.155716 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.158728 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.158855 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.164712 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555446-8n794"] Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.218621 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8dxf\" (UniqueName: \"kubernetes.io/projected/39c23ecc-c75b-450b-a8ff-351acf5384eb-kube-api-access-b8dxf\") pod \"auto-csr-approver-29555446-8n794\" (UID: \"39c23ecc-c75b-450b-a8ff-351acf5384eb\") " pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.321175 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8dxf\" (UniqueName: \"kubernetes.io/projected/39c23ecc-c75b-450b-a8ff-351acf5384eb-kube-api-access-b8dxf\") pod \"auto-csr-approver-29555446-8n794\" (UID: \"39c23ecc-c75b-450b-a8ff-351acf5384eb\") " pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.348038 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8dxf\" (UniqueName: \"kubernetes.io/projected/39c23ecc-c75b-450b-a8ff-351acf5384eb-kube-api-access-b8dxf\") pod \"auto-csr-approver-29555446-8n794\" (UID: \"39c23ecc-c75b-450b-a8ff-351acf5384eb\") " pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.475082 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:00 crc kubenswrapper[4778]: I0312 14:46:00.925115 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555446-8n794"] Mar 12 14:46:01 crc kubenswrapper[4778]: I0312 14:46:01.055512 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555446-8n794" event={"ID":"39c23ecc-c75b-450b-a8ff-351acf5384eb","Type":"ContainerStarted","Data":"595d06f338bb91adb2e450722d1ef8c57ba66b32da1700432f1c19563c6812ef"} Mar 12 14:46:03 crc kubenswrapper[4778]: I0312 14:46:03.074430 4778 generic.go:334] "Generic (PLEG): container finished" podID="39c23ecc-c75b-450b-a8ff-351acf5384eb" containerID="7fa3212d8016436bf15f0a3b9362ee13f653d69a236beda9f2c5ab4f28324438" exitCode=0 Mar 12 14:46:03 crc kubenswrapper[4778]: I0312 14:46:03.074490 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555446-8n794" event={"ID":"39c23ecc-c75b-450b-a8ff-351acf5384eb","Type":"ContainerDied","Data":"7fa3212d8016436bf15f0a3b9362ee13f653d69a236beda9f2c5ab4f28324438"} Mar 12 14:46:04 crc kubenswrapper[4778]: I0312 14:46:04.465165 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:04 crc kubenswrapper[4778]: I0312 14:46:04.501086 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8dxf\" (UniqueName: \"kubernetes.io/projected/39c23ecc-c75b-450b-a8ff-351acf5384eb-kube-api-access-b8dxf\") pod \"39c23ecc-c75b-450b-a8ff-351acf5384eb\" (UID: \"39c23ecc-c75b-450b-a8ff-351acf5384eb\") " Mar 12 14:46:04 crc kubenswrapper[4778]: I0312 14:46:04.509627 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39c23ecc-c75b-450b-a8ff-351acf5384eb-kube-api-access-b8dxf" (OuterVolumeSpecName: "kube-api-access-b8dxf") pod "39c23ecc-c75b-450b-a8ff-351acf5384eb" (UID: "39c23ecc-c75b-450b-a8ff-351acf5384eb"). InnerVolumeSpecName "kube-api-access-b8dxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:46:04 crc kubenswrapper[4778]: I0312 14:46:04.604117 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8dxf\" (UniqueName: \"kubernetes.io/projected/39c23ecc-c75b-450b-a8ff-351acf5384eb-kube-api-access-b8dxf\") on node \"crc\" DevicePath \"\"" Mar 12 14:46:05 crc kubenswrapper[4778]: I0312 14:46:05.093830 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555446-8n794" event={"ID":"39c23ecc-c75b-450b-a8ff-351acf5384eb","Type":"ContainerDied","Data":"595d06f338bb91adb2e450722d1ef8c57ba66b32da1700432f1c19563c6812ef"} Mar 12 14:46:05 crc kubenswrapper[4778]: I0312 14:46:05.094225 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="595d06f338bb91adb2e450722d1ef8c57ba66b32da1700432f1c19563c6812ef" Mar 12 14:46:05 crc kubenswrapper[4778]: I0312 14:46:05.093876 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555446-8n794" Mar 12 14:46:05 crc kubenswrapper[4778]: I0312 14:46:05.556169 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555440-prm5s"] Mar 12 14:46:05 crc kubenswrapper[4778]: I0312 14:46:05.568588 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555440-prm5s"] Mar 12 14:46:06 crc kubenswrapper[4778]: I0312 14:46:06.268839 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19be98d0-a8c0-4e30-926a-3ac799c6b576" path="/var/lib/kubelet/pods/19be98d0-a8c0-4e30-926a-3ac799c6b576/volumes" Mar 12 14:46:07 crc kubenswrapper[4778]: I0312 14:46:07.254068 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:46:07 crc kubenswrapper[4778]: E0312 14:46:07.254629 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:46:22 crc kubenswrapper[4778]: I0312 14:46:22.254529 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:46:22 crc kubenswrapper[4778]: E0312 14:46:22.255420 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:46:23 crc kubenswrapper[4778]: I0312 14:46:23.647211 4778 scope.go:117] "RemoveContainer" containerID="e11c1ed077d226d809927c1b3f1c2bde51c60c0a0093049ef2bd6c0eb78accb3" Mar 12 14:46:36 crc kubenswrapper[4778]: I0312 14:46:36.254825 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:46:36 crc kubenswrapper[4778]: E0312 14:46:36.255608 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.015906 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wd89s"] Mar 12 14:46:38 crc kubenswrapper[4778]: E0312 14:46:38.016576 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39c23ecc-c75b-450b-a8ff-351acf5384eb" containerName="oc" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.016592 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="39c23ecc-c75b-450b-a8ff-351acf5384eb" containerName="oc" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.016823 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="39c23ecc-c75b-450b-a8ff-351acf5384eb" containerName="oc" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.018617 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.038618 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd89s"] Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.116101 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-catalog-content\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.116225 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-utilities\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.116440 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt7jp\" (UniqueName: \"kubernetes.io/projected/55405211-a853-42e1-b97d-f876cfe8571e-kube-api-access-tt7jp\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.217738 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-utilities\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.218101 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt7jp\" (UniqueName: \"kubernetes.io/projected/55405211-a853-42e1-b97d-f876cfe8571e-kube-api-access-tt7jp\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.218361 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-catalog-content\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.218446 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-utilities\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.218688 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-catalog-content\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.241165 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt7jp\" (UniqueName: \"kubernetes.io/projected/55405211-a853-42e1-b97d-f876cfe8571e-kube-api-access-tt7jp\") pod \"community-operators-wd89s\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.338787 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:38 crc kubenswrapper[4778]: I0312 14:46:38.892097 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd89s"] Mar 12 14:46:39 crc kubenswrapper[4778]: I0312 14:46:39.383706 4778 generic.go:334] "Generic (PLEG): container finished" podID="55405211-a853-42e1-b97d-f876cfe8571e" containerID="e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f" exitCode=0 Mar 12 14:46:39 crc kubenswrapper[4778]: I0312 14:46:39.383926 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd89s" event={"ID":"55405211-a853-42e1-b97d-f876cfe8571e","Type":"ContainerDied","Data":"e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f"} Mar 12 14:46:39 crc kubenswrapper[4778]: I0312 14:46:39.383978 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd89s" event={"ID":"55405211-a853-42e1-b97d-f876cfe8571e","Type":"ContainerStarted","Data":"e21a0b422cf52b7acf33c1dc0980a93ccaf4f052a0f0d2e75d0729ce68636faa"} Mar 12 14:46:41 crc kubenswrapper[4778]: I0312 14:46:41.402158 4778 generic.go:334] "Generic (PLEG): container finished" podID="55405211-a853-42e1-b97d-f876cfe8571e" containerID="267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d" exitCode=0 Mar 12 14:46:41 crc kubenswrapper[4778]: I0312 14:46:41.402216 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd89s" event={"ID":"55405211-a853-42e1-b97d-f876cfe8571e","Type":"ContainerDied","Data":"267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d"} Mar 12 14:46:42 crc kubenswrapper[4778]: I0312 14:46:42.414288 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd89s" event={"ID":"55405211-a853-42e1-b97d-f876cfe8571e","Type":"ContainerStarted","Data":"66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174"} Mar 12 14:46:42 crc kubenswrapper[4778]: I0312 14:46:42.435326 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wd89s" podStartSLOduration=2.862101639 podStartE2EDuration="5.435311607s" podCreationTimestamp="2026-03-12 14:46:37 +0000 UTC" firstStartedPulling="2026-03-12 14:46:39.385927809 +0000 UTC m=+5817.834623205" lastFinishedPulling="2026-03-12 14:46:41.959137777 +0000 UTC m=+5820.407833173" observedRunningTime="2026-03-12 14:46:42.43083781 +0000 UTC m=+5820.879533216" watchObservedRunningTime="2026-03-12 14:46:42.435311607 +0000 UTC m=+5820.884007003" Mar 12 14:46:48 crc kubenswrapper[4778]: I0312 14:46:48.338933 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:48 crc kubenswrapper[4778]: I0312 14:46:48.339482 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:48 crc kubenswrapper[4778]: I0312 14:46:48.400291 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:48 crc kubenswrapper[4778]: I0312 14:46:48.580039 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:48 crc kubenswrapper[4778]: I0312 14:46:48.653580 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd89s"] Mar 12 14:46:49 crc kubenswrapper[4778]: I0312 14:46:49.253546 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:46:49 crc kubenswrapper[4778]: E0312 14:46:49.254024 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:46:50 crc kubenswrapper[4778]: I0312 14:46:50.480406 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wd89s" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="registry-server" containerID="cri-o://66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174" gracePeriod=2 Mar 12 14:46:50 crc kubenswrapper[4778]: I0312 14:46:50.946163 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.100826 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-catalog-content\") pod \"55405211-a853-42e1-b97d-f876cfe8571e\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.104489 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt7jp\" (UniqueName: \"kubernetes.io/projected/55405211-a853-42e1-b97d-f876cfe8571e-kube-api-access-tt7jp\") pod \"55405211-a853-42e1-b97d-f876cfe8571e\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.104663 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-utilities\") pod \"55405211-a853-42e1-b97d-f876cfe8571e\" (UID: \"55405211-a853-42e1-b97d-f876cfe8571e\") " Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.105278 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-utilities" (OuterVolumeSpecName: "utilities") pod "55405211-a853-42e1-b97d-f876cfe8571e" (UID: "55405211-a853-42e1-b97d-f876cfe8571e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.111419 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55405211-a853-42e1-b97d-f876cfe8571e-kube-api-access-tt7jp" (OuterVolumeSpecName: "kube-api-access-tt7jp") pod "55405211-a853-42e1-b97d-f876cfe8571e" (UID: "55405211-a853-42e1-b97d-f876cfe8571e"). InnerVolumeSpecName "kube-api-access-tt7jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.207586 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt7jp\" (UniqueName: \"kubernetes.io/projected/55405211-a853-42e1-b97d-f876cfe8571e-kube-api-access-tt7jp\") on node \"crc\" DevicePath \"\"" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.207626 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.322094 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55405211-a853-42e1-b97d-f876cfe8571e" (UID: "55405211-a853-42e1-b97d-f876cfe8571e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.412953 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55405211-a853-42e1-b97d-f876cfe8571e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.492527 4778 generic.go:334] "Generic (PLEG): container finished" podID="55405211-a853-42e1-b97d-f876cfe8571e" containerID="66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174" exitCode=0 Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.492576 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd89s" event={"ID":"55405211-a853-42e1-b97d-f876cfe8571e","Type":"ContainerDied","Data":"66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174"} Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.492611 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd89s" event={"ID":"55405211-a853-42e1-b97d-f876cfe8571e","Type":"ContainerDied","Data":"e21a0b422cf52b7acf33c1dc0980a93ccaf4f052a0f0d2e75d0729ce68636faa"} Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.492639 4778 scope.go:117] "RemoveContainer" containerID="66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.492799 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd89s" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.515170 4778 scope.go:117] "RemoveContainer" containerID="267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.541576 4778 scope.go:117] "RemoveContainer" containerID="e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.545920 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd89s"] Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.558858 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wd89s"] Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.591083 4778 scope.go:117] "RemoveContainer" containerID="66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174" Mar 12 14:46:51 crc kubenswrapper[4778]: E0312 14:46:51.591808 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174\": container with ID starting with 66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174 not found: ID does not exist" containerID="66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.591849 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174"} err="failed to get container status \"66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174\": rpc error: code = NotFound desc = could not find container \"66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174\": container with ID starting with 66a2cd8cc923470afb472e47dd40e5182ac80ecfec264a4d0f93ac2f50632174 not found: ID does not exist" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.591878 4778 scope.go:117] "RemoveContainer" containerID="267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d" Mar 12 14:46:51 crc kubenswrapper[4778]: E0312 14:46:51.592098 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d\": container with ID starting with 267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d not found: ID does not exist" containerID="267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.592124 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d"} err="failed to get container status \"267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d\": rpc error: code = NotFound desc = could not find container \"267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d\": container with ID starting with 267c4937eac5a95221a52aee7d8893da6636a3f2179067570591e31e2d3e940d not found: ID does not exist" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.592138 4778 scope.go:117] "RemoveContainer" containerID="e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f" Mar 12 14:46:51 crc kubenswrapper[4778]: E0312 14:46:51.592544 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f\": container with ID starting with e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f not found: ID does not exist" containerID="e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f" Mar 12 14:46:51 crc kubenswrapper[4778]: I0312 14:46:51.592574 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f"} err="failed to get container status \"e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f\": rpc error: code = NotFound desc = could not find container \"e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f\": container with ID starting with e5e5916972ff38e811789e88f3b01f57b526d86d6bbc87acdd5636aadde5a15f not found: ID does not exist" Mar 12 14:46:52 crc kubenswrapper[4778]: I0312 14:46:52.273369 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55405211-a853-42e1-b97d-f876cfe8571e" path="/var/lib/kubelet/pods/55405211-a853-42e1-b97d-f876cfe8571e/volumes" Mar 12 14:47:00 crc kubenswrapper[4778]: I0312 14:47:00.254612 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:47:00 crc kubenswrapper[4778]: E0312 14:47:00.255565 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:47:12 crc kubenswrapper[4778]: I0312 14:47:12.261485 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:47:12 crc kubenswrapper[4778]: E0312 14:47:12.262747 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.366683 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fsnnv"] Mar 12 14:47:13 crc kubenswrapper[4778]: E0312 14:47:13.367301 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="extract-utilities" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.367325 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="extract-utilities" Mar 12 14:47:13 crc kubenswrapper[4778]: E0312 14:47:13.367378 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="extract-content" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.367391 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="extract-content" Mar 12 14:47:13 crc kubenswrapper[4778]: E0312 14:47:13.367416 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="registry-server" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.367430 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="registry-server" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.367732 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="55405211-a853-42e1-b97d-f876cfe8571e" containerName="registry-server" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.369923 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.389919 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fsnnv"] Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.446321 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-catalog-content\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.446598 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfhbb\" (UniqueName: \"kubernetes.io/projected/26c2f0ef-3580-429e-8fc5-3ece50c4f023-kube-api-access-rfhbb\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.446862 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-utilities\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.549107 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-utilities\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.549568 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-catalog-content\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.549689 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfhbb\" (UniqueName: \"kubernetes.io/projected/26c2f0ef-3580-429e-8fc5-3ece50c4f023-kube-api-access-rfhbb\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.549569 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-utilities\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.549958 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-catalog-content\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.569250 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfhbb\" (UniqueName: \"kubernetes.io/projected/26c2f0ef-3580-429e-8fc5-3ece50c4f023-kube-api-access-rfhbb\") pod \"certified-operators-fsnnv\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:13 crc kubenswrapper[4778]: I0312 14:47:13.714482 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:14 crc kubenswrapper[4778]: I0312 14:47:14.198716 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fsnnv"] Mar 12 14:47:14 crc kubenswrapper[4778]: I0312 14:47:14.703097 4778 generic.go:334] "Generic (PLEG): container finished" podID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerID="231eb24c5f060d0284f56a7b47085c3333eabe57f70491b0a128a5966c6669b1" exitCode=0 Mar 12 14:47:14 crc kubenswrapper[4778]: I0312 14:47:14.703142 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerDied","Data":"231eb24c5f060d0284f56a7b47085c3333eabe57f70491b0a128a5966c6669b1"} Mar 12 14:47:14 crc kubenswrapper[4778]: I0312 14:47:14.703431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerStarted","Data":"8b1acbfd6f53aab1c6cf025f822a3ebe1171651a9a7c34804adfad48d82717bc"} Mar 12 14:47:14 crc kubenswrapper[4778]: I0312 14:47:14.705040 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:47:15 crc kubenswrapper[4778]: I0312 14:47:15.724869 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerStarted","Data":"b2b21cf4b6fbb46f6ba37f27234667ec7d7d7258895ed6cd412a23268a2bbaca"} Mar 12 14:47:17 crc kubenswrapper[4778]: I0312 14:47:17.754130 4778 generic.go:334] "Generic (PLEG): container finished" podID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerID="b2b21cf4b6fbb46f6ba37f27234667ec7d7d7258895ed6cd412a23268a2bbaca" exitCode=0 Mar 12 14:47:17 crc kubenswrapper[4778]: I0312 14:47:17.754207 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerDied","Data":"b2b21cf4b6fbb46f6ba37f27234667ec7d7d7258895ed6cd412a23268a2bbaca"} Mar 12 14:47:18 crc kubenswrapper[4778]: I0312 14:47:18.764882 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerStarted","Data":"c72b4f2bff4ba063399bb11f18084a9908fb796531a42d291dab6f2ebe39f85b"} Mar 12 14:47:18 crc kubenswrapper[4778]: I0312 14:47:18.786128 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fsnnv" podStartSLOduration=2.322515064 podStartE2EDuration="5.786104129s" podCreationTimestamp="2026-03-12 14:47:13 +0000 UTC" firstStartedPulling="2026-03-12 14:47:14.704795809 +0000 UTC m=+5853.153491205" lastFinishedPulling="2026-03-12 14:47:18.168384864 +0000 UTC m=+5856.617080270" observedRunningTime="2026-03-12 14:47:18.780074887 +0000 UTC m=+5857.228770293" watchObservedRunningTime="2026-03-12 14:47:18.786104129 +0000 UTC m=+5857.234799525" Mar 12 14:47:23 crc kubenswrapper[4778]: I0312 14:47:23.254424 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:47:23 crc kubenswrapper[4778]: E0312 14:47:23.254939 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:47:23 crc kubenswrapper[4778]: I0312 14:47:23.714859 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:23 crc kubenswrapper[4778]: I0312 14:47:23.716229 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:23 crc kubenswrapper[4778]: I0312 14:47:23.763670 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:23 crc kubenswrapper[4778]: I0312 14:47:23.882433 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:24 crc kubenswrapper[4778]: I0312 14:47:23.999983 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fsnnv"] Mar 12 14:47:25 crc kubenswrapper[4778]: I0312 14:47:25.830665 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fsnnv" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="registry-server" containerID="cri-o://c72b4f2bff4ba063399bb11f18084a9908fb796531a42d291dab6f2ebe39f85b" gracePeriod=2 Mar 12 14:47:26 crc kubenswrapper[4778]: I0312 14:47:26.847801 4778 generic.go:334] "Generic (PLEG): container finished" podID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerID="c72b4f2bff4ba063399bb11f18084a9908fb796531a42d291dab6f2ebe39f85b" exitCode=0 Mar 12 14:47:26 crc kubenswrapper[4778]: I0312 14:47:26.847937 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerDied","Data":"c72b4f2bff4ba063399bb11f18084a9908fb796531a42d291dab6f2ebe39f85b"} Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.581358 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.639557 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfhbb\" (UniqueName: \"kubernetes.io/projected/26c2f0ef-3580-429e-8fc5-3ece50c4f023-kube-api-access-rfhbb\") pod \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.639636 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-utilities\") pod \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.639750 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-catalog-content\") pod \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\" (UID: \"26c2f0ef-3580-429e-8fc5-3ece50c4f023\") " Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.640643 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-utilities" (OuterVolumeSpecName: "utilities") pod "26c2f0ef-3580-429e-8fc5-3ece50c4f023" (UID: "26c2f0ef-3580-429e-8fc5-3ece50c4f023"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.645713 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26c2f0ef-3580-429e-8fc5-3ece50c4f023-kube-api-access-rfhbb" (OuterVolumeSpecName: "kube-api-access-rfhbb") pod "26c2f0ef-3580-429e-8fc5-3ece50c4f023" (UID: "26c2f0ef-3580-429e-8fc5-3ece50c4f023"). InnerVolumeSpecName "kube-api-access-rfhbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.697632 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26c2f0ef-3580-429e-8fc5-3ece50c4f023" (UID: "26c2f0ef-3580-429e-8fc5-3ece50c4f023"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.742521 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfhbb\" (UniqueName: \"kubernetes.io/projected/26c2f0ef-3580-429e-8fc5-3ece50c4f023-kube-api-access-rfhbb\") on node \"crc\" DevicePath \"\"" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.742555 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.742564 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26c2f0ef-3580-429e-8fc5-3ece50c4f023-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.859856 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsnnv" event={"ID":"26c2f0ef-3580-429e-8fc5-3ece50c4f023","Type":"ContainerDied","Data":"8b1acbfd6f53aab1c6cf025f822a3ebe1171651a9a7c34804adfad48d82717bc"} Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.859906 4778 scope.go:117] "RemoveContainer" containerID="c72b4f2bff4ba063399bb11f18084a9908fb796531a42d291dab6f2ebe39f85b" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.860036 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsnnv" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.885129 4778 scope.go:117] "RemoveContainer" containerID="b2b21cf4b6fbb46f6ba37f27234667ec7d7d7258895ed6cd412a23268a2bbaca" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.908039 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fsnnv"] Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.910963 4778 scope.go:117] "RemoveContainer" containerID="231eb24c5f060d0284f56a7b47085c3333eabe57f70491b0a128a5966c6669b1" Mar 12 14:47:27 crc kubenswrapper[4778]: I0312 14:47:27.926936 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fsnnv"] Mar 12 14:47:28 crc kubenswrapper[4778]: I0312 14:47:28.264165 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" path="/var/lib/kubelet/pods/26c2f0ef-3580-429e-8fc5-3ece50c4f023/volumes" Mar 12 14:47:37 crc kubenswrapper[4778]: I0312 14:47:37.254480 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:47:37 crc kubenswrapper[4778]: E0312 14:47:37.255317 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:47:50 crc kubenswrapper[4778]: I0312 14:47:50.254992 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:47:50 crc kubenswrapper[4778]: E0312 14:47:50.256108 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.154331 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555448-g9lqj"] Mar 12 14:48:00 crc kubenswrapper[4778]: E0312 14:48:00.155391 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="extract-utilities" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.155407 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="extract-utilities" Mar 12 14:48:00 crc kubenswrapper[4778]: E0312 14:48:00.155451 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="registry-server" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.155460 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="registry-server" Mar 12 14:48:00 crc kubenswrapper[4778]: E0312 14:48:00.155483 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="extract-content" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.155491 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="extract-content" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.155716 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="26c2f0ef-3580-429e-8fc5-3ece50c4f023" containerName="registry-server" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.156592 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.159769 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.159994 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.160145 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.167254 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555448-g9lqj"] Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.296438 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqb64\" (UniqueName: \"kubernetes.io/projected/53a65a66-de4e-413e-a175-4d12db4e3f26-kube-api-access-zqb64\") pod \"auto-csr-approver-29555448-g9lqj\" (UID: \"53a65a66-de4e-413e-a175-4d12db4e3f26\") " pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.398770 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqb64\" (UniqueName: \"kubernetes.io/projected/53a65a66-de4e-413e-a175-4d12db4e3f26-kube-api-access-zqb64\") pod \"auto-csr-approver-29555448-g9lqj\" (UID: \"53a65a66-de4e-413e-a175-4d12db4e3f26\") " pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.428318 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqb64\" (UniqueName: \"kubernetes.io/projected/53a65a66-de4e-413e-a175-4d12db4e3f26-kube-api-access-zqb64\") pod \"auto-csr-approver-29555448-g9lqj\" (UID: \"53a65a66-de4e-413e-a175-4d12db4e3f26\") " pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.488607 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:00 crc kubenswrapper[4778]: W0312 14:48:00.959655 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53a65a66_de4e_413e_a175_4d12db4e3f26.slice/crio-c4c03bcf2cbf592eb2e1d01dd3fcd63207b161e20b0137bceffe7a12716b1809 WatchSource:0}: Error finding container c4c03bcf2cbf592eb2e1d01dd3fcd63207b161e20b0137bceffe7a12716b1809: Status 404 returned error can't find the container with id c4c03bcf2cbf592eb2e1d01dd3fcd63207b161e20b0137bceffe7a12716b1809 Mar 12 14:48:00 crc kubenswrapper[4778]: I0312 14:48:00.963096 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555448-g9lqj"] Mar 12 14:48:01 crc kubenswrapper[4778]: I0312 14:48:01.178587 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" event={"ID":"53a65a66-de4e-413e-a175-4d12db4e3f26","Type":"ContainerStarted","Data":"c4c03bcf2cbf592eb2e1d01dd3fcd63207b161e20b0137bceffe7a12716b1809"} Mar 12 14:48:03 crc kubenswrapper[4778]: I0312 14:48:03.200403 4778 generic.go:334] "Generic (PLEG): container finished" podID="53a65a66-de4e-413e-a175-4d12db4e3f26" containerID="26a11a81934702ff4eaece8862eb99dd5a6954a851baea01b2b49d973eba34bc" exitCode=0 Mar 12 14:48:03 crc kubenswrapper[4778]: I0312 14:48:03.200465 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" event={"ID":"53a65a66-de4e-413e-a175-4d12db4e3f26","Type":"ContainerDied","Data":"26a11a81934702ff4eaece8862eb99dd5a6954a851baea01b2b49d973eba34bc"} Mar 12 14:48:03 crc kubenswrapper[4778]: I0312 14:48:03.254604 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:48:03 crc kubenswrapper[4778]: E0312 14:48:03.255403 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:48:04 crc kubenswrapper[4778]: I0312 14:48:04.610333 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:04 crc kubenswrapper[4778]: I0312 14:48:04.791614 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqb64\" (UniqueName: \"kubernetes.io/projected/53a65a66-de4e-413e-a175-4d12db4e3f26-kube-api-access-zqb64\") pod \"53a65a66-de4e-413e-a175-4d12db4e3f26\" (UID: \"53a65a66-de4e-413e-a175-4d12db4e3f26\") " Mar 12 14:48:04 crc kubenswrapper[4778]: I0312 14:48:04.807420 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a65a66-de4e-413e-a175-4d12db4e3f26-kube-api-access-zqb64" (OuterVolumeSpecName: "kube-api-access-zqb64") pod "53a65a66-de4e-413e-a175-4d12db4e3f26" (UID: "53a65a66-de4e-413e-a175-4d12db4e3f26"). InnerVolumeSpecName "kube-api-access-zqb64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:48:04 crc kubenswrapper[4778]: I0312 14:48:04.894435 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqb64\" (UniqueName: \"kubernetes.io/projected/53a65a66-de4e-413e-a175-4d12db4e3f26-kube-api-access-zqb64\") on node \"crc\" DevicePath \"\"" Mar 12 14:48:05 crc kubenswrapper[4778]: I0312 14:48:05.217840 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" event={"ID":"53a65a66-de4e-413e-a175-4d12db4e3f26","Type":"ContainerDied","Data":"c4c03bcf2cbf592eb2e1d01dd3fcd63207b161e20b0137bceffe7a12716b1809"} Mar 12 14:48:05 crc kubenswrapper[4778]: I0312 14:48:05.218139 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4c03bcf2cbf592eb2e1d01dd3fcd63207b161e20b0137bceffe7a12716b1809" Mar 12 14:48:05 crc kubenswrapper[4778]: I0312 14:48:05.218068 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555448-g9lqj" Mar 12 14:48:05 crc kubenswrapper[4778]: I0312 14:48:05.687090 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555442-c58f5"] Mar 12 14:48:05 crc kubenswrapper[4778]: I0312 14:48:05.695315 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555442-c58f5"] Mar 12 14:48:06 crc kubenswrapper[4778]: I0312 14:48:06.264794 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86de30c2-8699-4966-8c1b-da67cdacae42" path="/var/lib/kubelet/pods/86de30c2-8699-4966-8c1b-da67cdacae42/volumes" Mar 12 14:48:14 crc kubenswrapper[4778]: I0312 14:48:14.254053 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:48:14 crc kubenswrapper[4778]: E0312 14:48:14.254820 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:48:23 crc kubenswrapper[4778]: I0312 14:48:23.757786 4778 scope.go:117] "RemoveContainer" containerID="629d070304f0ca91f60ef09f2871ae160406fbd39c685129feb137e8e63e7888" Mar 12 14:48:28 crc kubenswrapper[4778]: I0312 14:48:28.254447 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:48:28 crc kubenswrapper[4778]: E0312 14:48:28.255447 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:48:40 crc kubenswrapper[4778]: I0312 14:48:40.254555 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:48:40 crc kubenswrapper[4778]: E0312 14:48:40.255412 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:48:53 crc kubenswrapper[4778]: I0312 14:48:53.255035 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:48:53 crc kubenswrapper[4778]: E0312 14:48:53.256360 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:49:04 crc kubenswrapper[4778]: I0312 14:49:04.254708 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:49:04 crc kubenswrapper[4778]: E0312 14:49:04.255708 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:49:16 crc kubenswrapper[4778]: I0312 14:49:16.254396 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:49:16 crc kubenswrapper[4778]: E0312 14:49:16.255424 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:49:29 crc kubenswrapper[4778]: I0312 14:49:29.260744 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:49:29 crc kubenswrapper[4778]: E0312 14:49:29.261904 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:49:43 crc kubenswrapper[4778]: I0312 14:49:43.255027 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:49:43 crc kubenswrapper[4778]: E0312 14:49:43.259360 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:49:55 crc kubenswrapper[4778]: I0312 14:49:55.254849 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:49:55 crc kubenswrapper[4778]: E0312 14:49:55.256299 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.175237 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555450-wng5r"] Mar 12 14:50:00 crc kubenswrapper[4778]: E0312 14:50:00.176136 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a65a66-de4e-413e-a175-4d12db4e3f26" containerName="oc" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.176152 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a65a66-de4e-413e-a175-4d12db4e3f26" containerName="oc" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.176359 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a65a66-de4e-413e-a175-4d12db4e3f26" containerName="oc" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.176994 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.181010 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.181071 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.181010 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.190367 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555450-wng5r"] Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.358669 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44dqg\" (UniqueName: \"kubernetes.io/projected/58bc575b-de62-42e8-8393-0cdebe8a1ec7-kube-api-access-44dqg\") pod \"auto-csr-approver-29555450-wng5r\" (UID: \"58bc575b-de62-42e8-8393-0cdebe8a1ec7\") " pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.461588 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44dqg\" (UniqueName: \"kubernetes.io/projected/58bc575b-de62-42e8-8393-0cdebe8a1ec7-kube-api-access-44dqg\") pod \"auto-csr-approver-29555450-wng5r\" (UID: \"58bc575b-de62-42e8-8393-0cdebe8a1ec7\") " pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.485120 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44dqg\" (UniqueName: \"kubernetes.io/projected/58bc575b-de62-42e8-8393-0cdebe8a1ec7-kube-api-access-44dqg\") pod \"auto-csr-approver-29555450-wng5r\" (UID: \"58bc575b-de62-42e8-8393-0cdebe8a1ec7\") " pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:00 crc kubenswrapper[4778]: I0312 14:50:00.507522 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:01 crc kubenswrapper[4778]: I0312 14:50:01.029309 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555450-wng5r"] Mar 12 14:50:01 crc kubenswrapper[4778]: I0312 14:50:01.427057 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555450-wng5r" event={"ID":"58bc575b-de62-42e8-8393-0cdebe8a1ec7","Type":"ContainerStarted","Data":"9ed451c9d70977425643f13fc74daea81c277d539cb67ad214466aceb2d38ff6"} Mar 12 14:50:03 crc kubenswrapper[4778]: I0312 14:50:03.458458 4778 generic.go:334] "Generic (PLEG): container finished" podID="58bc575b-de62-42e8-8393-0cdebe8a1ec7" containerID="b673bf4baccca3cbc88953e0302f6d44002e09551d5876af0fec26f563392bf0" exitCode=0 Mar 12 14:50:03 crc kubenswrapper[4778]: I0312 14:50:03.458709 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555450-wng5r" event={"ID":"58bc575b-de62-42e8-8393-0cdebe8a1ec7","Type":"ContainerDied","Data":"b673bf4baccca3cbc88953e0302f6d44002e09551d5876af0fec26f563392bf0"} Mar 12 14:50:04 crc kubenswrapper[4778]: I0312 14:50:04.881132 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:04 crc kubenswrapper[4778]: I0312 14:50:04.953823 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44dqg\" (UniqueName: \"kubernetes.io/projected/58bc575b-de62-42e8-8393-0cdebe8a1ec7-kube-api-access-44dqg\") pod \"58bc575b-de62-42e8-8393-0cdebe8a1ec7\" (UID: \"58bc575b-de62-42e8-8393-0cdebe8a1ec7\") " Mar 12 14:50:04 crc kubenswrapper[4778]: I0312 14:50:04.959564 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58bc575b-de62-42e8-8393-0cdebe8a1ec7-kube-api-access-44dqg" (OuterVolumeSpecName: "kube-api-access-44dqg") pod "58bc575b-de62-42e8-8393-0cdebe8a1ec7" (UID: "58bc575b-de62-42e8-8393-0cdebe8a1ec7"). InnerVolumeSpecName "kube-api-access-44dqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:50:05 crc kubenswrapper[4778]: I0312 14:50:05.056508 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44dqg\" (UniqueName: \"kubernetes.io/projected/58bc575b-de62-42e8-8393-0cdebe8a1ec7-kube-api-access-44dqg\") on node \"crc\" DevicePath \"\"" Mar 12 14:50:05 crc kubenswrapper[4778]: I0312 14:50:05.476527 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555450-wng5r" event={"ID":"58bc575b-de62-42e8-8393-0cdebe8a1ec7","Type":"ContainerDied","Data":"9ed451c9d70977425643f13fc74daea81c277d539cb67ad214466aceb2d38ff6"} Mar 12 14:50:05 crc kubenswrapper[4778]: I0312 14:50:05.476570 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ed451c9d70977425643f13fc74daea81c277d539cb67ad214466aceb2d38ff6" Mar 12 14:50:05 crc kubenswrapper[4778]: I0312 14:50:05.476597 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555450-wng5r" Mar 12 14:50:05 crc kubenswrapper[4778]: I0312 14:50:05.954707 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555444-qlvql"] Mar 12 14:50:05 crc kubenswrapper[4778]: I0312 14:50:05.966173 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555444-qlvql"] Mar 12 14:50:06 crc kubenswrapper[4778]: I0312 14:50:06.254936 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:50:06 crc kubenswrapper[4778]: E0312 14:50:06.255756 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:50:06 crc kubenswrapper[4778]: I0312 14:50:06.273900 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0da8caed-335b-4730-a69b-b724585893f0" path="/var/lib/kubelet/pods/0da8caed-335b-4730-a69b-b724585893f0/volumes" Mar 12 14:50:21 crc kubenswrapper[4778]: I0312 14:50:21.254650 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:50:21 crc kubenswrapper[4778]: E0312 14:50:21.255815 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:50:23 crc kubenswrapper[4778]: I0312 14:50:23.921073 4778 scope.go:117] "RemoveContainer" containerID="1c045d93602ba35174db559893247fb3da6916ea0a208015b77cd0f47781b091" Mar 12 14:50:32 crc kubenswrapper[4778]: I0312 14:50:32.263551 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:50:32 crc kubenswrapper[4778]: I0312 14:50:32.757428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"f9dc5323f20567a96d1ddcd61f28e57c1fb446407246116e9b85f41f7b862a79"} Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.187356 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555452-crwhx"] Mar 12 14:52:00 crc kubenswrapper[4778]: E0312 14:52:00.188224 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58bc575b-de62-42e8-8393-0cdebe8a1ec7" containerName="oc" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.188237 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="58bc575b-de62-42e8-8393-0cdebe8a1ec7" containerName="oc" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.188437 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="58bc575b-de62-42e8-8393-0cdebe8a1ec7" containerName="oc" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.189036 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.194268 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.194300 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.195123 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.198012 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555452-crwhx"] Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.256507 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg2pl\" (UniqueName: \"kubernetes.io/projected/db3a8d80-262e-4c92-b07a-dcff65e0cd47-kube-api-access-bg2pl\") pod \"auto-csr-approver-29555452-crwhx\" (UID: \"db3a8d80-262e-4c92-b07a-dcff65e0cd47\") " pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.357815 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg2pl\" (UniqueName: \"kubernetes.io/projected/db3a8d80-262e-4c92-b07a-dcff65e0cd47-kube-api-access-bg2pl\") pod \"auto-csr-approver-29555452-crwhx\" (UID: \"db3a8d80-262e-4c92-b07a-dcff65e0cd47\") " pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.379596 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg2pl\" (UniqueName: \"kubernetes.io/projected/db3a8d80-262e-4c92-b07a-dcff65e0cd47-kube-api-access-bg2pl\") pod \"auto-csr-approver-29555452-crwhx\" (UID: \"db3a8d80-262e-4c92-b07a-dcff65e0cd47\") " pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:00 crc kubenswrapper[4778]: I0312 14:52:00.506917 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:01 crc kubenswrapper[4778]: I0312 14:52:01.010091 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555452-crwhx"] Mar 12 14:52:01 crc kubenswrapper[4778]: I0312 14:52:01.667715 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555452-crwhx" event={"ID":"db3a8d80-262e-4c92-b07a-dcff65e0cd47","Type":"ContainerStarted","Data":"f38c768e364d8f115a87d0bf5c94ef1dff3ea8fb8be912ee295bb78ac0faa144"} Mar 12 14:52:02 crc kubenswrapper[4778]: I0312 14:52:02.682076 4778 generic.go:334] "Generic (PLEG): container finished" podID="db3a8d80-262e-4c92-b07a-dcff65e0cd47" containerID="2e201785308313aa155d17696c3a92cd860cbcfcbc51f75878f68248fd82d5d8" exitCode=0 Mar 12 14:52:02 crc kubenswrapper[4778]: I0312 14:52:02.682145 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555452-crwhx" event={"ID":"db3a8d80-262e-4c92-b07a-dcff65e0cd47","Type":"ContainerDied","Data":"2e201785308313aa155d17696c3a92cd860cbcfcbc51f75878f68248fd82d5d8"} Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.090311 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.157344 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg2pl\" (UniqueName: \"kubernetes.io/projected/db3a8d80-262e-4c92-b07a-dcff65e0cd47-kube-api-access-bg2pl\") pod \"db3a8d80-262e-4c92-b07a-dcff65e0cd47\" (UID: \"db3a8d80-262e-4c92-b07a-dcff65e0cd47\") " Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.163551 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3a8d80-262e-4c92-b07a-dcff65e0cd47-kube-api-access-bg2pl" (OuterVolumeSpecName: "kube-api-access-bg2pl") pod "db3a8d80-262e-4c92-b07a-dcff65e0cd47" (UID: "db3a8d80-262e-4c92-b07a-dcff65e0cd47"). InnerVolumeSpecName "kube-api-access-bg2pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.259648 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg2pl\" (UniqueName: \"kubernetes.io/projected/db3a8d80-262e-4c92-b07a-dcff65e0cd47-kube-api-access-bg2pl\") on node \"crc\" DevicePath \"\"" Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.710554 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555452-crwhx" event={"ID":"db3a8d80-262e-4c92-b07a-dcff65e0cd47","Type":"ContainerDied","Data":"f38c768e364d8f115a87d0bf5c94ef1dff3ea8fb8be912ee295bb78ac0faa144"} Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.710628 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f38c768e364d8f115a87d0bf5c94ef1dff3ea8fb8be912ee295bb78ac0faa144" Mar 12 14:52:04 crc kubenswrapper[4778]: I0312 14:52:04.710658 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555452-crwhx" Mar 12 14:52:05 crc kubenswrapper[4778]: I0312 14:52:05.169743 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555446-8n794"] Mar 12 14:52:05 crc kubenswrapper[4778]: I0312 14:52:05.179631 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555446-8n794"] Mar 12 14:52:06 crc kubenswrapper[4778]: I0312 14:52:06.274691 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39c23ecc-c75b-450b-a8ff-351acf5384eb" path="/var/lib/kubelet/pods/39c23ecc-c75b-450b-a8ff-351acf5384eb/volumes" Mar 12 14:52:24 crc kubenswrapper[4778]: I0312 14:52:24.066541 4778 scope.go:117] "RemoveContainer" containerID="7fa3212d8016436bf15f0a3b9362ee13f653d69a236beda9f2c5ab4f28324438" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.362762 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5ksjs"] Mar 12 14:52:33 crc kubenswrapper[4778]: E0312 14:52:33.363866 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3a8d80-262e-4c92-b07a-dcff65e0cd47" containerName="oc" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.363882 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3a8d80-262e-4c92-b07a-dcff65e0cd47" containerName="oc" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.364131 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3a8d80-262e-4c92-b07a-dcff65e0cd47" containerName="oc" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.365707 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.382832 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ksjs"] Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.495962 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-catalog-content\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.496056 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-utilities\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.496542 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x49kn\" (UniqueName: \"kubernetes.io/projected/5ad61852-35ce-4f63-8876-d1231244f3a2-kube-api-access-x49kn\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.598183 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x49kn\" (UniqueName: \"kubernetes.io/projected/5ad61852-35ce-4f63-8876-d1231244f3a2-kube-api-access-x49kn\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.598499 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-catalog-content\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.598566 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-utilities\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.599114 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-utilities\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.599215 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-catalog-content\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.619107 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x49kn\" (UniqueName: \"kubernetes.io/projected/5ad61852-35ce-4f63-8876-d1231244f3a2-kube-api-access-x49kn\") pod \"redhat-marketplace-5ksjs\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:33 crc kubenswrapper[4778]: I0312 14:52:33.687456 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:52:34 crc kubenswrapper[4778]: I0312 14:52:34.168674 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ksjs"] Mar 12 14:52:35 crc kubenswrapper[4778]: I0312 14:52:35.093851 4778 generic.go:334] "Generic (PLEG): container finished" podID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerID="a7c672d4b0c14852ce23d5decfb704e09ea74bef2e325221b4069fcc117c4976" exitCode=0 Mar 12 14:52:35 crc kubenswrapper[4778]: I0312 14:52:35.093920 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerDied","Data":"a7c672d4b0c14852ce23d5decfb704e09ea74bef2e325221b4069fcc117c4976"} Mar 12 14:52:35 crc kubenswrapper[4778]: I0312 14:52:35.093985 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerStarted","Data":"7b0e362a57eae3105936688cad6d42fcb7298e40ea92f03730e3ac287d78abff"} Mar 12 14:52:35 crc kubenswrapper[4778]: I0312 14:52:35.097963 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:52:36 crc kubenswrapper[4778]: I0312 14:52:36.105328 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerStarted","Data":"6a4c01059a32908d762eeffa04e27aa1d1e32ba2e0e5fac8ef80086f88864c0c"} Mar 12 14:52:38 crc kubenswrapper[4778]: I0312 14:52:38.032264 4778 generic.go:334] "Generic (PLEG): container finished" podID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerID="6a4c01059a32908d762eeffa04e27aa1d1e32ba2e0e5fac8ef80086f88864c0c" exitCode=0 Mar 12 14:52:38 crc kubenswrapper[4778]: I0312 14:52:38.034806 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerDied","Data":"6a4c01059a32908d762eeffa04e27aa1d1e32ba2e0e5fac8ef80086f88864c0c"} Mar 12 14:52:38 crc kubenswrapper[4778]: E0312 14:52:38.153615 4778 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad: Requesting bearer token: invalid status code from registry 502 (Bad Gateway)" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad" Mar 12 14:52:38 crc kubenswrapper[4778]: E0312 14:52:38.153791 4778 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad,Command:[/bin/opm],Args:[serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOMEMLIMIT,Value:20MiB,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x49kn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5ksjs_openshift-marketplace(5ad61852-35ce-4f63-8876-d1231244f3a2): ErrImagePull: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad: Requesting bearer token: invalid status code from registry 502 (Bad Gateway)" logger="UnhandledError" Mar 12 14:52:38 crc kubenswrapper[4778]: E0312 14:52:38.156101 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad: Requesting bearer token: invalid status code from registry 502 (Bad Gateway)\"" pod="openshift-marketplace/redhat-marketplace-5ksjs" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" Mar 12 14:52:39 crc kubenswrapper[4778]: E0312 14:52:39.051526 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"\"" pod="openshift-marketplace/redhat-marketplace-5ksjs" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" Mar 12 14:52:55 crc kubenswrapper[4778]: I0312 14:52:55.225224 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerStarted","Data":"a07c96607ff2d89770d88eed7cd8da0599a0a5a80c5f7751bcfe6d3fc67c632e"} Mar 12 14:52:55 crc kubenswrapper[4778]: I0312 14:52:55.261249 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5ksjs" podStartSLOduration=2.631164154 podStartE2EDuration="22.261222032s" podCreationTimestamp="2026-03-12 14:52:33 +0000 UTC" firstStartedPulling="2026-03-12 14:52:35.09763692 +0000 UTC m=+6173.546332326" lastFinishedPulling="2026-03-12 14:52:54.727694778 +0000 UTC m=+6193.176390204" observedRunningTime="2026-03-12 14:52:55.250172978 +0000 UTC m=+6193.698868374" watchObservedRunningTime="2026-03-12 14:52:55.261222032 +0000 UTC m=+6193.709917458" Mar 12 14:52:58 crc kubenswrapper[4778]: I0312 14:52:58.557409 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:52:58 crc kubenswrapper[4778]: I0312 14:52:58.558044 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:53:03 crc kubenswrapper[4778]: I0312 14:53:03.688460 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:53:03 crc kubenswrapper[4778]: I0312 14:53:03.688792 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:53:03 crc kubenswrapper[4778]: I0312 14:53:03.776541 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:53:04 crc kubenswrapper[4778]: I0312 14:53:04.377113 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:53:07 crc kubenswrapper[4778]: I0312 14:53:07.923631 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ksjs"] Mar 12 14:53:07 crc kubenswrapper[4778]: I0312 14:53:07.928969 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5ksjs" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="registry-server" containerID="cri-o://a07c96607ff2d89770d88eed7cd8da0599a0a5a80c5f7751bcfe6d3fc67c632e" gracePeriod=2 Mar 12 14:53:08 crc kubenswrapper[4778]: I0312 14:53:08.918734 4778 generic.go:334] "Generic (PLEG): container finished" podID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerID="a07c96607ff2d89770d88eed7cd8da0599a0a5a80c5f7751bcfe6d3fc67c632e" exitCode=0 Mar 12 14:53:08 crc kubenswrapper[4778]: I0312 14:53:08.918776 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerDied","Data":"a07c96607ff2d89770d88eed7cd8da0599a0a5a80c5f7751bcfe6d3fc67c632e"} Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.137656 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.259499 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-catalog-content\") pod \"5ad61852-35ce-4f63-8876-d1231244f3a2\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.259554 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-utilities\") pod \"5ad61852-35ce-4f63-8876-d1231244f3a2\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.259633 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x49kn\" (UniqueName: \"kubernetes.io/projected/5ad61852-35ce-4f63-8876-d1231244f3a2-kube-api-access-x49kn\") pod \"5ad61852-35ce-4f63-8876-d1231244f3a2\" (UID: \"5ad61852-35ce-4f63-8876-d1231244f3a2\") " Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.260408 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-utilities" (OuterVolumeSpecName: "utilities") pod "5ad61852-35ce-4f63-8876-d1231244f3a2" (UID: "5ad61852-35ce-4f63-8876-d1231244f3a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.264947 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad61852-35ce-4f63-8876-d1231244f3a2-kube-api-access-x49kn" (OuterVolumeSpecName: "kube-api-access-x49kn") pod "5ad61852-35ce-4f63-8876-d1231244f3a2" (UID: "5ad61852-35ce-4f63-8876-d1231244f3a2"). InnerVolumeSpecName "kube-api-access-x49kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.289067 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ad61852-35ce-4f63-8876-d1231244f3a2" (UID: "5ad61852-35ce-4f63-8876-d1231244f3a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.362631 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.362667 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad61852-35ce-4f63-8876-d1231244f3a2-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.362698 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x49kn\" (UniqueName: \"kubernetes.io/projected/5ad61852-35ce-4f63-8876-d1231244f3a2-kube-api-access-x49kn\") on node \"crc\" DevicePath \"\"" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.933174 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ksjs" event={"ID":"5ad61852-35ce-4f63-8876-d1231244f3a2","Type":"ContainerDied","Data":"7b0e362a57eae3105936688cad6d42fcb7298e40ea92f03730e3ac287d78abff"} Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.933561 4778 scope.go:117] "RemoveContainer" containerID="a07c96607ff2d89770d88eed7cd8da0599a0a5a80c5f7751bcfe6d3fc67c632e" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.933261 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ksjs" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.967009 4778 scope.go:117] "RemoveContainer" containerID="6a4c01059a32908d762eeffa04e27aa1d1e32ba2e0e5fac8ef80086f88864c0c" Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.973718 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ksjs"] Mar 12 14:53:09 crc kubenswrapper[4778]: I0312 14:53:09.982080 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ksjs"] Mar 12 14:53:10 crc kubenswrapper[4778]: I0312 14:53:10.016777 4778 scope.go:117] "RemoveContainer" containerID="a7c672d4b0c14852ce23d5decfb704e09ea74bef2e325221b4069fcc117c4976" Mar 12 14:53:10 crc kubenswrapper[4778]: I0312 14:53:10.268648 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" path="/var/lib/kubelet/pods/5ad61852-35ce-4f63-8876-d1231244f3a2/volumes" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.079812 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7g7g"] Mar 12 14:53:25 crc kubenswrapper[4778]: E0312 14:53:25.081130 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="registry-server" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.081153 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="registry-server" Mar 12 14:53:25 crc kubenswrapper[4778]: E0312 14:53:25.081225 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="extract-utilities" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.081237 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="extract-utilities" Mar 12 14:53:25 crc kubenswrapper[4778]: E0312 14:53:25.081271 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="extract-content" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.081281 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="extract-content" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.081603 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad61852-35ce-4f63-8876-d1231244f3a2" containerName="registry-server" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.084095 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.106236 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7g7g"] Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.198295 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-utilities\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.199364 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-catalog-content\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.199811 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdtk5\" (UniqueName: \"kubernetes.io/projected/3a389bb4-d9de-42e4-911c-1a07358309b3-kube-api-access-fdtk5\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.302045 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-utilities\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.302139 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-catalog-content\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.302264 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdtk5\" (UniqueName: \"kubernetes.io/projected/3a389bb4-d9de-42e4-911c-1a07358309b3-kube-api-access-fdtk5\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.305943 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-utilities\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.310986 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-catalog-content\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.322882 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdtk5\" (UniqueName: \"kubernetes.io/projected/3a389bb4-d9de-42e4-911c-1a07358309b3-kube-api-access-fdtk5\") pod \"redhat-operators-x7g7g\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.406346 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:25 crc kubenswrapper[4778]: I0312 14:53:25.867137 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7g7g"] Mar 12 14:53:26 crc kubenswrapper[4778]: I0312 14:53:26.240559 4778 generic.go:334] "Generic (PLEG): container finished" podID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerID="9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73" exitCode=0 Mar 12 14:53:26 crc kubenswrapper[4778]: I0312 14:53:26.240618 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerDied","Data":"9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73"} Mar 12 14:53:26 crc kubenswrapper[4778]: I0312 14:53:26.240902 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerStarted","Data":"3f81bb08789404e82e902669b50736be3048d7835a6c22fd2dce8060ad7a7309"} Mar 12 14:53:28 crc kubenswrapper[4778]: I0312 14:53:28.557905 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:53:28 crc kubenswrapper[4778]: I0312 14:53:28.558616 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:53:29 crc kubenswrapper[4778]: I0312 14:53:29.268124 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerStarted","Data":"98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df"} Mar 12 14:53:38 crc kubenswrapper[4778]: I0312 14:53:38.352783 4778 generic.go:334] "Generic (PLEG): container finished" podID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerID="98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df" exitCode=0 Mar 12 14:53:38 crc kubenswrapper[4778]: I0312 14:53:38.353000 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerDied","Data":"98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df"} Mar 12 14:53:39 crc kubenswrapper[4778]: I0312 14:53:39.363708 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerStarted","Data":"b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200"} Mar 12 14:53:39 crc kubenswrapper[4778]: I0312 14:53:39.399910 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7g7g" podStartSLOduration=1.8473961829999999 podStartE2EDuration="14.399889708s" podCreationTimestamp="2026-03-12 14:53:25 +0000 UTC" firstStartedPulling="2026-03-12 14:53:26.242282794 +0000 UTC m=+6224.690978180" lastFinishedPulling="2026-03-12 14:53:38.794776309 +0000 UTC m=+6237.243471705" observedRunningTime="2026-03-12 14:53:39.397994974 +0000 UTC m=+6237.846690370" watchObservedRunningTime="2026-03-12 14:53:39.399889708 +0000 UTC m=+6237.848585104" Mar 12 14:53:45 crc kubenswrapper[4778]: I0312 14:53:45.406990 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:45 crc kubenswrapper[4778]: I0312 14:53:45.407622 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:53:46 crc kubenswrapper[4778]: I0312 14:53:46.462431 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x7g7g" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" probeResult="failure" output=< Mar 12 14:53:46 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:53:46 crc kubenswrapper[4778]: > Mar 12 14:53:56 crc kubenswrapper[4778]: I0312 14:53:56.451556 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x7g7g" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" probeResult="failure" output=< Mar 12 14:53:56 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:53:56 crc kubenswrapper[4778]: > Mar 12 14:53:58 crc kubenswrapper[4778]: I0312 14:53:58.558307 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:53:58 crc kubenswrapper[4778]: I0312 14:53:58.558631 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:53:58 crc kubenswrapper[4778]: I0312 14:53:58.558685 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:53:58 crc kubenswrapper[4778]: I0312 14:53:58.559458 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9dc5323f20567a96d1ddcd61f28e57c1fb446407246116e9b85f41f7b862a79"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:53:58 crc kubenswrapper[4778]: I0312 14:53:58.559510 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://f9dc5323f20567a96d1ddcd61f28e57c1fb446407246116e9b85f41f7b862a79" gracePeriod=600 Mar 12 14:53:59 crc kubenswrapper[4778]: I0312 14:53:59.535582 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="f9dc5323f20567a96d1ddcd61f28e57c1fb446407246116e9b85f41f7b862a79" exitCode=0 Mar 12 14:53:59 crc kubenswrapper[4778]: I0312 14:53:59.535836 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"f9dc5323f20567a96d1ddcd61f28e57c1fb446407246116e9b85f41f7b862a79"} Mar 12 14:53:59 crc kubenswrapper[4778]: I0312 14:53:59.536147 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3"} Mar 12 14:53:59 crc kubenswrapper[4778]: I0312 14:53:59.536172 4778 scope.go:117] "RemoveContainer" containerID="e714113346a3db81a8ab4456acd91be95b7042ec696820890f89fb14190436c4" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.144928 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555454-45jhq"] Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.147149 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.149142 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.149164 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.150222 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.159291 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555454-45jhq"] Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.249996 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l84cm\" (UniqueName: \"kubernetes.io/projected/3c838df0-ebc6-482a-8c4a-54e2650a121a-kube-api-access-l84cm\") pod \"auto-csr-approver-29555454-45jhq\" (UID: \"3c838df0-ebc6-482a-8c4a-54e2650a121a\") " pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.352372 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l84cm\" (UniqueName: \"kubernetes.io/projected/3c838df0-ebc6-482a-8c4a-54e2650a121a-kube-api-access-l84cm\") pod \"auto-csr-approver-29555454-45jhq\" (UID: \"3c838df0-ebc6-482a-8c4a-54e2650a121a\") " pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.374379 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l84cm\" (UniqueName: \"kubernetes.io/projected/3c838df0-ebc6-482a-8c4a-54e2650a121a-kube-api-access-l84cm\") pod \"auto-csr-approver-29555454-45jhq\" (UID: \"3c838df0-ebc6-482a-8c4a-54e2650a121a\") " pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:00 crc kubenswrapper[4778]: I0312 14:54:00.467436 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:01 crc kubenswrapper[4778]: I0312 14:54:01.003774 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555454-45jhq"] Mar 12 14:54:01 crc kubenswrapper[4778]: W0312 14:54:01.003802 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c838df0_ebc6_482a_8c4a_54e2650a121a.slice/crio-74a5f56be069b7c9eeb83716a3c1db85f0876c5d29e54cb7bf731b147dd7cb92 WatchSource:0}: Error finding container 74a5f56be069b7c9eeb83716a3c1db85f0876c5d29e54cb7bf731b147dd7cb92: Status 404 returned error can't find the container with id 74a5f56be069b7c9eeb83716a3c1db85f0876c5d29e54cb7bf731b147dd7cb92 Mar 12 14:54:01 crc kubenswrapper[4778]: I0312 14:54:01.566068 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555454-45jhq" event={"ID":"3c838df0-ebc6-482a-8c4a-54e2650a121a","Type":"ContainerStarted","Data":"74a5f56be069b7c9eeb83716a3c1db85f0876c5d29e54cb7bf731b147dd7cb92"} Mar 12 14:54:02 crc kubenswrapper[4778]: I0312 14:54:02.576280 4778 generic.go:334] "Generic (PLEG): container finished" podID="3c838df0-ebc6-482a-8c4a-54e2650a121a" containerID="a9ba0939d14aff103f5b46662e1f25d349a1d48a1eb2501077c92e0d8ad3aee1" exitCode=0 Mar 12 14:54:02 crc kubenswrapper[4778]: I0312 14:54:02.576353 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555454-45jhq" event={"ID":"3c838df0-ebc6-482a-8c4a-54e2650a121a","Type":"ContainerDied","Data":"a9ba0939d14aff103f5b46662e1f25d349a1d48a1eb2501077c92e0d8ad3aee1"} Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.105684 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.238599 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l84cm\" (UniqueName: \"kubernetes.io/projected/3c838df0-ebc6-482a-8c4a-54e2650a121a-kube-api-access-l84cm\") pod \"3c838df0-ebc6-482a-8c4a-54e2650a121a\" (UID: \"3c838df0-ebc6-482a-8c4a-54e2650a121a\") " Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.252226 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c838df0-ebc6-482a-8c4a-54e2650a121a-kube-api-access-l84cm" (OuterVolumeSpecName: "kube-api-access-l84cm") pod "3c838df0-ebc6-482a-8c4a-54e2650a121a" (UID: "3c838df0-ebc6-482a-8c4a-54e2650a121a"). InnerVolumeSpecName "kube-api-access-l84cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.343066 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l84cm\" (UniqueName: \"kubernetes.io/projected/3c838df0-ebc6-482a-8c4a-54e2650a121a-kube-api-access-l84cm\") on node \"crc\" DevicePath \"\"" Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.599077 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555454-45jhq" event={"ID":"3c838df0-ebc6-482a-8c4a-54e2650a121a","Type":"ContainerDied","Data":"74a5f56be069b7c9eeb83716a3c1db85f0876c5d29e54cb7bf731b147dd7cb92"} Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.599121 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74a5f56be069b7c9eeb83716a3c1db85f0876c5d29e54cb7bf731b147dd7cb92" Mar 12 14:54:04 crc kubenswrapper[4778]: I0312 14:54:04.599214 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555454-45jhq" Mar 12 14:54:05 crc kubenswrapper[4778]: I0312 14:54:05.188361 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555448-g9lqj"] Mar 12 14:54:05 crc kubenswrapper[4778]: I0312 14:54:05.195839 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555448-g9lqj"] Mar 12 14:54:06 crc kubenswrapper[4778]: I0312 14:54:06.272482 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a65a66-de4e-413e-a175-4d12db4e3f26" path="/var/lib/kubelet/pods/53a65a66-de4e-413e-a175-4d12db4e3f26/volumes" Mar 12 14:54:06 crc kubenswrapper[4778]: I0312 14:54:06.469895 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x7g7g" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" probeResult="failure" output=< Mar 12 14:54:06 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 14:54:06 crc kubenswrapper[4778]: > Mar 12 14:54:15 crc kubenswrapper[4778]: I0312 14:54:15.459557 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:54:15 crc kubenswrapper[4778]: I0312 14:54:15.523894 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:54:15 crc kubenswrapper[4778]: I0312 14:54:15.716144 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7g7g"] Mar 12 14:54:16 crc kubenswrapper[4778]: I0312 14:54:16.728295 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7g7g" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" containerID="cri-o://b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200" gracePeriod=2 Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.266599 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.316048 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-catalog-content\") pod \"3a389bb4-d9de-42e4-911c-1a07358309b3\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.316123 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdtk5\" (UniqueName: \"kubernetes.io/projected/3a389bb4-d9de-42e4-911c-1a07358309b3-kube-api-access-fdtk5\") pod \"3a389bb4-d9de-42e4-911c-1a07358309b3\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.316173 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-utilities\") pod \"3a389bb4-d9de-42e4-911c-1a07358309b3\" (UID: \"3a389bb4-d9de-42e4-911c-1a07358309b3\") " Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.317864 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-utilities" (OuterVolumeSpecName: "utilities") pod "3a389bb4-d9de-42e4-911c-1a07358309b3" (UID: "3a389bb4-d9de-42e4-911c-1a07358309b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.326544 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a389bb4-d9de-42e4-911c-1a07358309b3-kube-api-access-fdtk5" (OuterVolumeSpecName: "kube-api-access-fdtk5") pod "3a389bb4-d9de-42e4-911c-1a07358309b3" (UID: "3a389bb4-d9de-42e4-911c-1a07358309b3"). InnerVolumeSpecName "kube-api-access-fdtk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.419247 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdtk5\" (UniqueName: \"kubernetes.io/projected/3a389bb4-d9de-42e4-911c-1a07358309b3-kube-api-access-fdtk5\") on node \"crc\" DevicePath \"\"" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.419291 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.454124 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a389bb4-d9de-42e4-911c-1a07358309b3" (UID: "3a389bb4-d9de-42e4-911c-1a07358309b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.520877 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a389bb4-d9de-42e4-911c-1a07358309b3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.745329 4778 generic.go:334] "Generic (PLEG): container finished" podID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerID="b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200" exitCode=0 Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.745410 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7g7g" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.745440 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerDied","Data":"b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200"} Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.745825 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7g7g" event={"ID":"3a389bb4-d9de-42e4-911c-1a07358309b3","Type":"ContainerDied","Data":"3f81bb08789404e82e902669b50736be3048d7835a6c22fd2dce8060ad7a7309"} Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.745875 4778 scope.go:117] "RemoveContainer" containerID="b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.781805 4778 scope.go:117] "RemoveContainer" containerID="98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.805131 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7g7g"] Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.819784 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7g7g"] Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.849903 4778 scope.go:117] "RemoveContainer" containerID="9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.882327 4778 scope.go:117] "RemoveContainer" containerID="b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200" Mar 12 14:54:17 crc kubenswrapper[4778]: E0312 14:54:17.882917 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200\": container with ID starting with b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200 not found: ID does not exist" containerID="b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.882962 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200"} err="failed to get container status \"b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200\": rpc error: code = NotFound desc = could not find container \"b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200\": container with ID starting with b50b6859d478e73a4f11f0a0e6fc939da1da9c5141ae0e6e8fbde7c1628ea200 not found: ID does not exist" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.882995 4778 scope.go:117] "RemoveContainer" containerID="98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df" Mar 12 14:54:17 crc kubenswrapper[4778]: E0312 14:54:17.883800 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df\": container with ID starting with 98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df not found: ID does not exist" containerID="98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.883886 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df"} err="failed to get container status \"98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df\": rpc error: code = NotFound desc = could not find container \"98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df\": container with ID starting with 98ec30106f08de9786eaaf89b5f0ef4b60b88d5f013083bc3efa87082bbf30df not found: ID does not exist" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.883952 4778 scope.go:117] "RemoveContainer" containerID="9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73" Mar 12 14:54:17 crc kubenswrapper[4778]: E0312 14:54:17.884555 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73\": container with ID starting with 9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73 not found: ID does not exist" containerID="9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73" Mar 12 14:54:17 crc kubenswrapper[4778]: I0312 14:54:17.884599 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73"} err="failed to get container status \"9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73\": rpc error: code = NotFound desc = could not find container \"9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73\": container with ID starting with 9ed6fef4e5b3590168e0a6e17950a074fe47863fe6c84cee2e3d716bc4a91c73 not found: ID does not exist" Mar 12 14:54:18 crc kubenswrapper[4778]: I0312 14:54:18.265120 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" path="/var/lib/kubelet/pods/3a389bb4-d9de-42e4-911c-1a07358309b3/volumes" Mar 12 14:54:24 crc kubenswrapper[4778]: I0312 14:54:24.218000 4778 scope.go:117] "RemoveContainer" containerID="26a11a81934702ff4eaece8862eb99dd5a6954a851baea01b2b49d973eba34bc" Mar 12 14:55:58 crc kubenswrapper[4778]: I0312 14:55:58.557531 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:55:58 crc kubenswrapper[4778]: I0312 14:55:58.558071 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.146613 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555456-vtfp6"] Mar 12 14:56:00 crc kubenswrapper[4778]: E0312 14:56:00.147423 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="extract-utilities" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.147438 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="extract-utilities" Mar 12 14:56:00 crc kubenswrapper[4778]: E0312 14:56:00.147453 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="extract-content" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.147461 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="extract-content" Mar 12 14:56:00 crc kubenswrapper[4778]: E0312 14:56:00.147473 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.147481 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" Mar 12 14:56:00 crc kubenswrapper[4778]: E0312 14:56:00.147496 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c838df0-ebc6-482a-8c4a-54e2650a121a" containerName="oc" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.147503 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c838df0-ebc6-482a-8c4a-54e2650a121a" containerName="oc" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.147713 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c838df0-ebc6-482a-8c4a-54e2650a121a" containerName="oc" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.147735 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a389bb4-d9de-42e4-911c-1a07358309b3" containerName="registry-server" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.148453 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.152677 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.152998 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.153404 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.159848 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555456-vtfp6"] Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.289669 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp58k\" (UniqueName: \"kubernetes.io/projected/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7-kube-api-access-sp58k\") pod \"auto-csr-approver-29555456-vtfp6\" (UID: \"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7\") " pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.391820 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp58k\" (UniqueName: \"kubernetes.io/projected/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7-kube-api-access-sp58k\") pod \"auto-csr-approver-29555456-vtfp6\" (UID: \"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7\") " pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.408708 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp58k\" (UniqueName: \"kubernetes.io/projected/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7-kube-api-access-sp58k\") pod \"auto-csr-approver-29555456-vtfp6\" (UID: \"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7\") " pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.489224 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:00 crc kubenswrapper[4778]: I0312 14:56:00.957865 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555456-vtfp6"] Mar 12 14:56:00 crc kubenswrapper[4778]: W0312 14:56:00.964004 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc0add53_7611_4f91_bf0b_cf5fea5bb9d7.slice/crio-b0c4898d6bd30fc8c90f320f594da5113bb471287af8bcbfb72dd799afc629e0 WatchSource:0}: Error finding container b0c4898d6bd30fc8c90f320f594da5113bb471287af8bcbfb72dd799afc629e0: Status 404 returned error can't find the container with id b0c4898d6bd30fc8c90f320f594da5113bb471287af8bcbfb72dd799afc629e0 Mar 12 14:56:01 crc kubenswrapper[4778]: I0312 14:56:01.734984 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" event={"ID":"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7","Type":"ContainerStarted","Data":"b0c4898d6bd30fc8c90f320f594da5113bb471287af8bcbfb72dd799afc629e0"} Mar 12 14:56:04 crc kubenswrapper[4778]: I0312 14:56:04.796397 4778 generic.go:334] "Generic (PLEG): container finished" podID="fc0add53-7611-4f91-bf0b-cf5fea5bb9d7" containerID="d7d76c5b2f5b6d4767497e4e99746de9373b74f615023933a47cba956a1bacb0" exitCode=0 Mar 12 14:56:04 crc kubenswrapper[4778]: I0312 14:56:04.796471 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" event={"ID":"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7","Type":"ContainerDied","Data":"d7d76c5b2f5b6d4767497e4e99746de9373b74f615023933a47cba956a1bacb0"} Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.209607 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.304848 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp58k\" (UniqueName: \"kubernetes.io/projected/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7-kube-api-access-sp58k\") pod \"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7\" (UID: \"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7\") " Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.310507 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7-kube-api-access-sp58k" (OuterVolumeSpecName: "kube-api-access-sp58k") pod "fc0add53-7611-4f91-bf0b-cf5fea5bb9d7" (UID: "fc0add53-7611-4f91-bf0b-cf5fea5bb9d7"). InnerVolumeSpecName "kube-api-access-sp58k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.407598 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp58k\" (UniqueName: \"kubernetes.io/projected/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7-kube-api-access-sp58k\") on node \"crc\" DevicePath \"\"" Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.813868 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" event={"ID":"fc0add53-7611-4f91-bf0b-cf5fea5bb9d7","Type":"ContainerDied","Data":"b0c4898d6bd30fc8c90f320f594da5113bb471287af8bcbfb72dd799afc629e0"} Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.813914 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0c4898d6bd30fc8c90f320f594da5113bb471287af8bcbfb72dd799afc629e0" Mar 12 14:56:06 crc kubenswrapper[4778]: I0312 14:56:06.813943 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555456-vtfp6" Mar 12 14:56:07 crc kubenswrapper[4778]: I0312 14:56:07.284380 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555450-wng5r"] Mar 12 14:56:07 crc kubenswrapper[4778]: I0312 14:56:07.293457 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555450-wng5r"] Mar 12 14:56:08 crc kubenswrapper[4778]: I0312 14:56:08.265236 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58bc575b-de62-42e8-8393-0cdebe8a1ec7" path="/var/lib/kubelet/pods/58bc575b-de62-42e8-8393-0cdebe8a1ec7/volumes" Mar 12 14:56:24 crc kubenswrapper[4778]: I0312 14:56:24.323342 4778 scope.go:117] "RemoveContainer" containerID="b673bf4baccca3cbc88953e0302f6d44002e09551d5876af0fec26f563392bf0" Mar 12 14:56:28 crc kubenswrapper[4778]: I0312 14:56:28.558109 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:56:28 crc kubenswrapper[4778]: I0312 14:56:28.558672 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:56:58 crc kubenswrapper[4778]: I0312 14:56:58.558079 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 14:56:58 crc kubenswrapper[4778]: I0312 14:56:58.558889 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 14:56:58 crc kubenswrapper[4778]: I0312 14:56:58.558972 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 14:56:58 crc kubenswrapper[4778]: I0312 14:56:58.560077 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 14:56:58 crc kubenswrapper[4778]: I0312 14:56:58.560140 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" gracePeriod=600 Mar 12 14:56:58 crc kubenswrapper[4778]: E0312 14:56:58.684224 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:56:59 crc kubenswrapper[4778]: I0312 14:56:59.296884 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" exitCode=0 Mar 12 14:56:59 crc kubenswrapper[4778]: I0312 14:56:59.296978 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3"} Mar 12 14:56:59 crc kubenswrapper[4778]: I0312 14:56:59.297225 4778 scope.go:117] "RemoveContainer" containerID="f9dc5323f20567a96d1ddcd61f28e57c1fb446407246116e9b85f41f7b862a79" Mar 12 14:56:59 crc kubenswrapper[4778]: I0312 14:56:59.297867 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:56:59 crc kubenswrapper[4778]: E0312 14:56:59.298213 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:57:10 crc kubenswrapper[4778]: I0312 14:57:10.255017 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:57:10 crc kubenswrapper[4778]: E0312 14:57:10.256085 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:57:22 crc kubenswrapper[4778]: I0312 14:57:22.407816 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:57:22 crc kubenswrapper[4778]: E0312 14:57:22.408628 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:57:34 crc kubenswrapper[4778]: I0312 14:57:34.253832 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:57:34 crc kubenswrapper[4778]: E0312 14:57:34.254645 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:57:48 crc kubenswrapper[4778]: I0312 14:57:48.254362 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:57:48 crc kubenswrapper[4778]: E0312 14:57:48.255139 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.774573 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h6mpk"] Mar 12 14:57:55 crc kubenswrapper[4778]: E0312 14:57:55.776264 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0add53-7611-4f91-bf0b-cf5fea5bb9d7" containerName="oc" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.776295 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0add53-7611-4f91-bf0b-cf5fea5bb9d7" containerName="oc" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.776735 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0add53-7611-4f91-bf0b-cf5fea5bb9d7" containerName="oc" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.779342 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.792791 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6mpk"] Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.921874 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-catalog-content\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.921955 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smnsf\" (UniqueName: \"kubernetes.io/projected/51704535-7590-40f1-8114-59f5032b1c86-kube-api-access-smnsf\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:55 crc kubenswrapper[4778]: I0312 14:57:55.922198 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-utilities\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.024207 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-catalog-content\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.024273 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smnsf\" (UniqueName: \"kubernetes.io/projected/51704535-7590-40f1-8114-59f5032b1c86-kube-api-access-smnsf\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.024378 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-utilities\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.024834 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-catalog-content\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.024904 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-utilities\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.047968 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smnsf\" (UniqueName: \"kubernetes.io/projected/51704535-7590-40f1-8114-59f5032b1c86-kube-api-access-smnsf\") pod \"certified-operators-h6mpk\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.112277 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.582279 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6mpk"] Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.911775 4778 generic.go:334] "Generic (PLEG): container finished" podID="51704535-7590-40f1-8114-59f5032b1c86" containerID="b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e" exitCode=0 Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.911822 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerDied","Data":"b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e"} Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.911846 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerStarted","Data":"4b7a7e0582fa56ba0004e62c821c92769140c78141a5756002ed142e7685d7aa"} Mar 12 14:57:56 crc kubenswrapper[4778]: I0312 14:57:56.914119 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 14:57:57 crc kubenswrapper[4778]: I0312 14:57:57.925283 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerStarted","Data":"5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835"} Mar 12 14:57:59 crc kubenswrapper[4778]: E0312 14:57:59.269809 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51704535_7590_40f1_8114_59f5032b1c86.slice/crio-5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835.scope\": RecentStats: unable to find data in memory cache]" Mar 12 14:57:59 crc kubenswrapper[4778]: I0312 14:57:59.947713 4778 generic.go:334] "Generic (PLEG): container finished" podID="51704535-7590-40f1-8114-59f5032b1c86" containerID="5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835" exitCode=0 Mar 12 14:57:59 crc kubenswrapper[4778]: I0312 14:57:59.947822 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerDied","Data":"5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835"} Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.151853 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555458-2kqth"] Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.154410 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.157590 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.158149 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.158599 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.162316 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555458-2kqth"] Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.230820 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp75b\" (UniqueName: \"kubernetes.io/projected/0fa9dd73-2656-43b3-a6cb-634d312a166e-kube-api-access-kp75b\") pod \"auto-csr-approver-29555458-2kqth\" (UID: \"0fa9dd73-2656-43b3-a6cb-634d312a166e\") " pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.333052 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp75b\" (UniqueName: \"kubernetes.io/projected/0fa9dd73-2656-43b3-a6cb-634d312a166e-kube-api-access-kp75b\") pod \"auto-csr-approver-29555458-2kqth\" (UID: \"0fa9dd73-2656-43b3-a6cb-634d312a166e\") " pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.366989 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp75b\" (UniqueName: \"kubernetes.io/projected/0fa9dd73-2656-43b3-a6cb-634d312a166e-kube-api-access-kp75b\") pod \"auto-csr-approver-29555458-2kqth\" (UID: \"0fa9dd73-2656-43b3-a6cb-634d312a166e\") " pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.477936 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.930507 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555458-2kqth"] Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.960123 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerStarted","Data":"3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59"} Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.961633 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555458-2kqth" event={"ID":"0fa9dd73-2656-43b3-a6cb-634d312a166e","Type":"ContainerStarted","Data":"51039247d67b2c4a116ad3728b96577d34ccb5f1f60aba63583d2068ea1b0883"} Mar 12 14:58:00 crc kubenswrapper[4778]: I0312 14:58:00.997357 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h6mpk" podStartSLOduration=2.5274186480000003 podStartE2EDuration="5.997341823s" podCreationTimestamp="2026-03-12 14:57:55 +0000 UTC" firstStartedPulling="2026-03-12 14:57:56.913813567 +0000 UTC m=+6495.362508963" lastFinishedPulling="2026-03-12 14:58:00.383736742 +0000 UTC m=+6498.832432138" observedRunningTime="2026-03-12 14:58:00.988604085 +0000 UTC m=+6499.437299481" watchObservedRunningTime="2026-03-12 14:58:00.997341823 +0000 UTC m=+6499.446037219" Mar 12 14:58:03 crc kubenswrapper[4778]: I0312 14:58:03.254543 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:58:03 crc kubenswrapper[4778]: E0312 14:58:03.255278 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:58:03 crc kubenswrapper[4778]: I0312 14:58:03.993660 4778 generic.go:334] "Generic (PLEG): container finished" podID="0fa9dd73-2656-43b3-a6cb-634d312a166e" containerID="a5663c78d0886a072205a20f2510ea67c65b15026159b43c8bf3ff0037ce7434" exitCode=0 Mar 12 14:58:03 crc kubenswrapper[4778]: I0312 14:58:03.993821 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555458-2kqth" event={"ID":"0fa9dd73-2656-43b3-a6cb-634d312a166e","Type":"ContainerDied","Data":"a5663c78d0886a072205a20f2510ea67c65b15026159b43c8bf3ff0037ce7434"} Mar 12 14:58:05 crc kubenswrapper[4778]: I0312 14:58:05.367751 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:05 crc kubenswrapper[4778]: I0312 14:58:05.383479 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp75b\" (UniqueName: \"kubernetes.io/projected/0fa9dd73-2656-43b3-a6cb-634d312a166e-kube-api-access-kp75b\") pod \"0fa9dd73-2656-43b3-a6cb-634d312a166e\" (UID: \"0fa9dd73-2656-43b3-a6cb-634d312a166e\") " Mar 12 14:58:05 crc kubenswrapper[4778]: I0312 14:58:05.391513 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa9dd73-2656-43b3-a6cb-634d312a166e-kube-api-access-kp75b" (OuterVolumeSpecName: "kube-api-access-kp75b") pod "0fa9dd73-2656-43b3-a6cb-634d312a166e" (UID: "0fa9dd73-2656-43b3-a6cb-634d312a166e"). InnerVolumeSpecName "kube-api-access-kp75b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:58:05 crc kubenswrapper[4778]: I0312 14:58:05.484770 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp75b\" (UniqueName: \"kubernetes.io/projected/0fa9dd73-2656-43b3-a6cb-634d312a166e-kube-api-access-kp75b\") on node \"crc\" DevicePath \"\"" Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.025998 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555458-2kqth" event={"ID":"0fa9dd73-2656-43b3-a6cb-634d312a166e","Type":"ContainerDied","Data":"51039247d67b2c4a116ad3728b96577d34ccb5f1f60aba63583d2068ea1b0883"} Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.026059 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51039247d67b2c4a116ad3728b96577d34ccb5f1f60aba63583d2068ea1b0883" Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.026108 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555458-2kqth" Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.113687 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.115308 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.160566 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.434560 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555452-crwhx"] Mar 12 14:58:06 crc kubenswrapper[4778]: I0312 14:58:06.446651 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555452-crwhx"] Mar 12 14:58:07 crc kubenswrapper[4778]: I0312 14:58:07.091085 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:58:07 crc kubenswrapper[4778]: I0312 14:58:07.146797 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h6mpk"] Mar 12 14:58:08 crc kubenswrapper[4778]: I0312 14:58:08.267088 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db3a8d80-262e-4c92-b07a-dcff65e0cd47" path="/var/lib/kubelet/pods/db3a8d80-262e-4c92-b07a-dcff65e0cd47/volumes" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.052130 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h6mpk" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="registry-server" containerID="cri-o://3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59" gracePeriod=2 Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.645998 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.773075 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-utilities\") pod \"51704535-7590-40f1-8114-59f5032b1c86\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.773132 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-catalog-content\") pod \"51704535-7590-40f1-8114-59f5032b1c86\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.773234 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smnsf\" (UniqueName: \"kubernetes.io/projected/51704535-7590-40f1-8114-59f5032b1c86-kube-api-access-smnsf\") pod \"51704535-7590-40f1-8114-59f5032b1c86\" (UID: \"51704535-7590-40f1-8114-59f5032b1c86\") " Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.773845 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-utilities" (OuterVolumeSpecName: "utilities") pod "51704535-7590-40f1-8114-59f5032b1c86" (UID: "51704535-7590-40f1-8114-59f5032b1c86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.781226 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51704535-7590-40f1-8114-59f5032b1c86-kube-api-access-smnsf" (OuterVolumeSpecName: "kube-api-access-smnsf") pod "51704535-7590-40f1-8114-59f5032b1c86" (UID: "51704535-7590-40f1-8114-59f5032b1c86"). InnerVolumeSpecName "kube-api-access-smnsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.849812 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51704535-7590-40f1-8114-59f5032b1c86" (UID: "51704535-7590-40f1-8114-59f5032b1c86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.875452 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.875487 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51704535-7590-40f1-8114-59f5032b1c86-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 14:58:09 crc kubenswrapper[4778]: I0312 14:58:09.875523 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smnsf\" (UniqueName: \"kubernetes.io/projected/51704535-7590-40f1-8114-59f5032b1c86-kube-api-access-smnsf\") on node \"crc\" DevicePath \"\"" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.062061 4778 generic.go:334] "Generic (PLEG): container finished" podID="51704535-7590-40f1-8114-59f5032b1c86" containerID="3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59" exitCode=0 Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.062132 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerDied","Data":"3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59"} Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.062156 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6mpk" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.062209 4778 scope.go:117] "RemoveContainer" containerID="3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.062349 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6mpk" event={"ID":"51704535-7590-40f1-8114-59f5032b1c86","Type":"ContainerDied","Data":"4b7a7e0582fa56ba0004e62c821c92769140c78141a5756002ed142e7685d7aa"} Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.090152 4778 scope.go:117] "RemoveContainer" containerID="5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.108599 4778 scope.go:117] "RemoveContainer" containerID="b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.112525 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h6mpk"] Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.121927 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h6mpk"] Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.150111 4778 scope.go:117] "RemoveContainer" containerID="3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59" Mar 12 14:58:10 crc kubenswrapper[4778]: E0312 14:58:10.150533 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59\": container with ID starting with 3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59 not found: ID does not exist" containerID="3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.150562 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59"} err="failed to get container status \"3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59\": rpc error: code = NotFound desc = could not find container \"3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59\": container with ID starting with 3a32f3e5d5134f200ee666a165f28c8a9bd2278b33b760b45d99e2265913bf59 not found: ID does not exist" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.150584 4778 scope.go:117] "RemoveContainer" containerID="5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835" Mar 12 14:58:10 crc kubenswrapper[4778]: E0312 14:58:10.150886 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835\": container with ID starting with 5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835 not found: ID does not exist" containerID="5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.150905 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835"} err="failed to get container status \"5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835\": rpc error: code = NotFound desc = could not find container \"5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835\": container with ID starting with 5c1140e2a3da443934a38be5417957e3a9e964721156d718e7e3eeb6dc137835 not found: ID does not exist" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.150917 4778 scope.go:117] "RemoveContainer" containerID="b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e" Mar 12 14:58:10 crc kubenswrapper[4778]: E0312 14:58:10.151209 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e\": container with ID starting with b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e not found: ID does not exist" containerID="b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.151265 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e"} err="failed to get container status \"b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e\": rpc error: code = NotFound desc = could not find container \"b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e\": container with ID starting with b419a433a52be5384840da5dbbba31c34114c87a32568aa1691306f3eab3966e not found: ID does not exist" Mar 12 14:58:10 crc kubenswrapper[4778]: I0312 14:58:10.266242 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51704535-7590-40f1-8114-59f5032b1c86" path="/var/lib/kubelet/pods/51704535-7590-40f1-8114-59f5032b1c86/volumes" Mar 12 14:58:18 crc kubenswrapper[4778]: I0312 14:58:18.253990 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:58:18 crc kubenswrapper[4778]: E0312 14:58:18.254478 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:58:24 crc kubenswrapper[4778]: I0312 14:58:24.419855 4778 scope.go:117] "RemoveContainer" containerID="2e201785308313aa155d17696c3a92cd860cbcfcbc51f75878f68248fd82d5d8" Mar 12 14:58:30 crc kubenswrapper[4778]: I0312 14:58:30.254973 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:58:30 crc kubenswrapper[4778]: E0312 14:58:30.256475 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:58:42 crc kubenswrapper[4778]: I0312 14:58:42.262287 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:58:42 crc kubenswrapper[4778]: E0312 14:58:42.263554 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:58:53 crc kubenswrapper[4778]: I0312 14:58:53.253986 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:58:53 crc kubenswrapper[4778]: E0312 14:58:53.254870 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:59:08 crc kubenswrapper[4778]: I0312 14:59:08.254418 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:59:08 crc kubenswrapper[4778]: E0312 14:59:08.255146 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:59:19 crc kubenswrapper[4778]: I0312 14:59:19.254774 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:59:19 crc kubenswrapper[4778]: E0312 14:59:19.255434 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:59:32 crc kubenswrapper[4778]: I0312 14:59:32.266598 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:59:32 crc kubenswrapper[4778]: E0312 14:59:32.267501 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:59:43 crc kubenswrapper[4778]: I0312 14:59:43.254130 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:59:43 crc kubenswrapper[4778]: E0312 14:59:43.254916 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 14:59:55 crc kubenswrapper[4778]: I0312 14:59:55.254437 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 14:59:55 crc kubenswrapper[4778]: E0312 14:59:55.255088 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.155283 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555460-6bwr2"] Mar 12 15:00:00 crc kubenswrapper[4778]: E0312 15:00:00.156341 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="registry-server" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.156357 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="registry-server" Mar 12 15:00:00 crc kubenswrapper[4778]: E0312 15:00:00.156367 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="extract-content" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.156375 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="extract-content" Mar 12 15:00:00 crc kubenswrapper[4778]: E0312 15:00:00.156390 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa9dd73-2656-43b3-a6cb-634d312a166e" containerName="oc" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.156397 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa9dd73-2656-43b3-a6cb-634d312a166e" containerName="oc" Mar 12 15:00:00 crc kubenswrapper[4778]: E0312 15:00:00.156416 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="extract-utilities" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.156425 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="extract-utilities" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.156676 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa9dd73-2656-43b3-a6cb-634d312a166e" containerName="oc" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.156704 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="51704535-7590-40f1-8114-59f5032b1c86" containerName="registry-server" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.157498 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.160262 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.160469 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.160856 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.178486 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq"] Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.179969 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.184221 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.184447 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.208230 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555460-6bwr2"] Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.216314 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq"] Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.301617 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faaa79f5-e391-4347-a4c3-c0a63518f540-secret-volume\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.301699 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faaa79f5-e391-4347-a4c3-c0a63518f540-config-volume\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.302346 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5lx6\" (UniqueName: \"kubernetes.io/projected/7baca351-722e-4d7e-972e-04513fae6e0b-kube-api-access-t5lx6\") pod \"auto-csr-approver-29555460-6bwr2\" (UID: \"7baca351-722e-4d7e-972e-04513fae6e0b\") " pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.302456 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjwlt\" (UniqueName: \"kubernetes.io/projected/faaa79f5-e391-4347-a4c3-c0a63518f540-kube-api-access-wjwlt\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.404214 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faaa79f5-e391-4347-a4c3-c0a63518f540-secret-volume\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.404273 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faaa79f5-e391-4347-a4c3-c0a63518f540-config-volume\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.404344 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5lx6\" (UniqueName: \"kubernetes.io/projected/7baca351-722e-4d7e-972e-04513fae6e0b-kube-api-access-t5lx6\") pod \"auto-csr-approver-29555460-6bwr2\" (UID: \"7baca351-722e-4d7e-972e-04513fae6e0b\") " pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.404505 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjwlt\" (UniqueName: \"kubernetes.io/projected/faaa79f5-e391-4347-a4c3-c0a63518f540-kube-api-access-wjwlt\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.406914 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faaa79f5-e391-4347-a4c3-c0a63518f540-config-volume\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.420659 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faaa79f5-e391-4347-a4c3-c0a63518f540-secret-volume\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.437502 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjwlt\" (UniqueName: \"kubernetes.io/projected/faaa79f5-e391-4347-a4c3-c0a63518f540-kube-api-access-wjwlt\") pod \"collect-profiles-29555460-t9qnq\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.437972 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5lx6\" (UniqueName: \"kubernetes.io/projected/7baca351-722e-4d7e-972e-04513fae6e0b-kube-api-access-t5lx6\") pod \"auto-csr-approver-29555460-6bwr2\" (UID: \"7baca351-722e-4d7e-972e-04513fae6e0b\") " pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.493108 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.506694 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:00 crc kubenswrapper[4778]: I0312 15:00:00.952862 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq"] Mar 12 15:00:01 crc kubenswrapper[4778]: I0312 15:00:01.038802 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555460-6bwr2"] Mar 12 15:00:01 crc kubenswrapper[4778]: W0312 15:00:01.053712 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7baca351_722e_4d7e_972e_04513fae6e0b.slice/crio-2b91083877674451481cea3ea78cfa01e88aafaff731fd1bd24142b45296aabd WatchSource:0}: Error finding container 2b91083877674451481cea3ea78cfa01e88aafaff731fd1bd24142b45296aabd: Status 404 returned error can't find the container with id 2b91083877674451481cea3ea78cfa01e88aafaff731fd1bd24142b45296aabd Mar 12 15:00:01 crc kubenswrapper[4778]: I0312 15:00:01.129268 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" event={"ID":"faaa79f5-e391-4347-a4c3-c0a63518f540","Type":"ContainerStarted","Data":"cc08a56b84f3ce3e074748729850a071bc430c114346dec2c3743a62ed94931b"} Mar 12 15:00:01 crc kubenswrapper[4778]: I0312 15:00:01.130598 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" event={"ID":"7baca351-722e-4d7e-972e-04513fae6e0b","Type":"ContainerStarted","Data":"2b91083877674451481cea3ea78cfa01e88aafaff731fd1bd24142b45296aabd"} Mar 12 15:00:02 crc kubenswrapper[4778]: I0312 15:00:02.140326 4778 generic.go:334] "Generic (PLEG): container finished" podID="faaa79f5-e391-4347-a4c3-c0a63518f540" containerID="3fe64b13554004be3fbf12b211482af1a85c6f10472ba77c6e1462c0d628fd9a" exitCode=0 Mar 12 15:00:02 crc kubenswrapper[4778]: I0312 15:00:02.140401 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" event={"ID":"faaa79f5-e391-4347-a4c3-c0a63518f540","Type":"ContainerDied","Data":"3fe64b13554004be3fbf12b211482af1a85c6f10472ba77c6e1462c0d628fd9a"} Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.506970 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.562741 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faaa79f5-e391-4347-a4c3-c0a63518f540-config-volume\") pod \"faaa79f5-e391-4347-a4c3-c0a63518f540\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.562829 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faaa79f5-e391-4347-a4c3-c0a63518f540-secret-volume\") pod \"faaa79f5-e391-4347-a4c3-c0a63518f540\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.562873 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjwlt\" (UniqueName: \"kubernetes.io/projected/faaa79f5-e391-4347-a4c3-c0a63518f540-kube-api-access-wjwlt\") pod \"faaa79f5-e391-4347-a4c3-c0a63518f540\" (UID: \"faaa79f5-e391-4347-a4c3-c0a63518f540\") " Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.563659 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faaa79f5-e391-4347-a4c3-c0a63518f540-config-volume" (OuterVolumeSpecName: "config-volume") pod "faaa79f5-e391-4347-a4c3-c0a63518f540" (UID: "faaa79f5-e391-4347-a4c3-c0a63518f540"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.568971 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faaa79f5-e391-4347-a4c3-c0a63518f540-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "faaa79f5-e391-4347-a4c3-c0a63518f540" (UID: "faaa79f5-e391-4347-a4c3-c0a63518f540"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.577456 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faaa79f5-e391-4347-a4c3-c0a63518f540-kube-api-access-wjwlt" (OuterVolumeSpecName: "kube-api-access-wjwlt") pod "faaa79f5-e391-4347-a4c3-c0a63518f540" (UID: "faaa79f5-e391-4347-a4c3-c0a63518f540"). InnerVolumeSpecName "kube-api-access-wjwlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.665658 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faaa79f5-e391-4347-a4c3-c0a63518f540-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.665708 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faaa79f5-e391-4347-a4c3-c0a63518f540-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 15:00:03 crc kubenswrapper[4778]: I0312 15:00:03.665728 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjwlt\" (UniqueName: \"kubernetes.io/projected/faaa79f5-e391-4347-a4c3-c0a63518f540-kube-api-access-wjwlt\") on node \"crc\" DevicePath \"\"" Mar 12 15:00:04 crc kubenswrapper[4778]: I0312 15:00:04.159465 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" Mar 12 15:00:04 crc kubenswrapper[4778]: I0312 15:00:04.159567 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555460-t9qnq" event={"ID":"faaa79f5-e391-4347-a4c3-c0a63518f540","Type":"ContainerDied","Data":"cc08a56b84f3ce3e074748729850a071bc430c114346dec2c3743a62ed94931b"} Mar 12 15:00:04 crc kubenswrapper[4778]: I0312 15:00:04.159869 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc08a56b84f3ce3e074748729850a071bc430c114346dec2c3743a62ed94931b" Mar 12 15:00:04 crc kubenswrapper[4778]: I0312 15:00:04.579761 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r"] Mar 12 15:00:04 crc kubenswrapper[4778]: I0312 15:00:04.588485 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555415-jjk6r"] Mar 12 15:00:05 crc kubenswrapper[4778]: I0312 15:00:05.169084 4778 generic.go:334] "Generic (PLEG): container finished" podID="7baca351-722e-4d7e-972e-04513fae6e0b" containerID="64150eeb0f1f171e7d11ada7712192a8c533967a0e598d41c325a6422f027d7a" exitCode=0 Mar 12 15:00:05 crc kubenswrapper[4778]: I0312 15:00:05.169367 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" event={"ID":"7baca351-722e-4d7e-972e-04513fae6e0b","Type":"ContainerDied","Data":"64150eeb0f1f171e7d11ada7712192a8c533967a0e598d41c325a6422f027d7a"} Mar 12 15:00:06 crc kubenswrapper[4778]: I0312 15:00:06.281443 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c6027ea-d1ed-4df0-bbe7-6904d2722fbc" path="/var/lib/kubelet/pods/9c6027ea-d1ed-4df0-bbe7-6904d2722fbc/volumes" Mar 12 15:00:06 crc kubenswrapper[4778]: I0312 15:00:06.540104 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:06 crc kubenswrapper[4778]: I0312 15:00:06.626536 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5lx6\" (UniqueName: \"kubernetes.io/projected/7baca351-722e-4d7e-972e-04513fae6e0b-kube-api-access-t5lx6\") pod \"7baca351-722e-4d7e-972e-04513fae6e0b\" (UID: \"7baca351-722e-4d7e-972e-04513fae6e0b\") " Mar 12 15:00:06 crc kubenswrapper[4778]: I0312 15:00:06.631817 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7baca351-722e-4d7e-972e-04513fae6e0b-kube-api-access-t5lx6" (OuterVolumeSpecName: "kube-api-access-t5lx6") pod "7baca351-722e-4d7e-972e-04513fae6e0b" (UID: "7baca351-722e-4d7e-972e-04513fae6e0b"). InnerVolumeSpecName "kube-api-access-t5lx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:00:06 crc kubenswrapper[4778]: I0312 15:00:06.727794 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5lx6\" (UniqueName: \"kubernetes.io/projected/7baca351-722e-4d7e-972e-04513fae6e0b-kube-api-access-t5lx6\") on node \"crc\" DevicePath \"\"" Mar 12 15:00:07 crc kubenswrapper[4778]: I0312 15:00:07.193246 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" event={"ID":"7baca351-722e-4d7e-972e-04513fae6e0b","Type":"ContainerDied","Data":"2b91083877674451481cea3ea78cfa01e88aafaff731fd1bd24142b45296aabd"} Mar 12 15:00:07 crc kubenswrapper[4778]: I0312 15:00:07.193285 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555460-6bwr2" Mar 12 15:00:07 crc kubenswrapper[4778]: I0312 15:00:07.193289 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b91083877674451481cea3ea78cfa01e88aafaff731fd1bd24142b45296aabd" Mar 12 15:00:07 crc kubenswrapper[4778]: I0312 15:00:07.597322 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555454-45jhq"] Mar 12 15:00:07 crc kubenswrapper[4778]: I0312 15:00:07.604845 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555454-45jhq"] Mar 12 15:00:08 crc kubenswrapper[4778]: I0312 15:00:08.254548 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:00:08 crc kubenswrapper[4778]: E0312 15:00:08.254845 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:00:08 crc kubenswrapper[4778]: I0312 15:00:08.264119 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c838df0-ebc6-482a-8c4a-54e2650a121a" path="/var/lib/kubelet/pods/3c838df0-ebc6-482a-8c4a-54e2650a121a/volumes" Mar 12 15:00:23 crc kubenswrapper[4778]: I0312 15:00:23.254875 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:00:23 crc kubenswrapper[4778]: E0312 15:00:23.256007 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:00:24 crc kubenswrapper[4778]: I0312 15:00:24.537992 4778 scope.go:117] "RemoveContainer" containerID="a9ba0939d14aff103f5b46662e1f25d349a1d48a1eb2501077c92e0d8ad3aee1" Mar 12 15:00:24 crc kubenswrapper[4778]: I0312 15:00:24.611250 4778 scope.go:117] "RemoveContainer" containerID="e47d44b34f9f52eb0c1249aedb361a64e96dcc50294b7036054124a9fc860b25" Mar 12 15:00:35 crc kubenswrapper[4778]: I0312 15:00:35.254646 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:00:35 crc kubenswrapper[4778]: E0312 15:00:35.255476 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:00:46 crc kubenswrapper[4778]: I0312 15:00:46.254999 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:00:46 crc kubenswrapper[4778]: E0312 15:00:46.256137 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:00:57 crc kubenswrapper[4778]: I0312 15:00:57.254835 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:00:57 crc kubenswrapper[4778]: E0312 15:00:57.255739 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.157482 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29555461-lmqk9"] Mar 12 15:01:00 crc kubenswrapper[4778]: E0312 15:01:00.158745 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faaa79f5-e391-4347-a4c3-c0a63518f540" containerName="collect-profiles" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.158760 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="faaa79f5-e391-4347-a4c3-c0a63518f540" containerName="collect-profiles" Mar 12 15:01:00 crc kubenswrapper[4778]: E0312 15:01:00.158780 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7baca351-722e-4d7e-972e-04513fae6e0b" containerName="oc" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.158786 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="7baca351-722e-4d7e-972e-04513fae6e0b" containerName="oc" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.158949 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="7baca351-722e-4d7e-972e-04513fae6e0b" containerName="oc" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.158970 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="faaa79f5-e391-4347-a4c3-c0a63518f540" containerName="collect-profiles" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.159583 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.178801 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-fernet-keys\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.178974 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-combined-ca-bundle\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.179013 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lls6t\" (UniqueName: \"kubernetes.io/projected/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-kube-api-access-lls6t\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.179112 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-config-data\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.226912 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29555461-lmqk9"] Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.280583 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lls6t\" (UniqueName: \"kubernetes.io/projected/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-kube-api-access-lls6t\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.281004 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-config-data\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.281277 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-fernet-keys\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.281682 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-combined-ca-bundle\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.286949 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-fernet-keys\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.287704 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-config-data\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.288530 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-combined-ca-bundle\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.304862 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lls6t\" (UniqueName: \"kubernetes.io/projected/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-kube-api-access-lls6t\") pod \"keystone-cron-29555461-lmqk9\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.481044 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:00 crc kubenswrapper[4778]: I0312 15:01:00.967277 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29555461-lmqk9"] Mar 12 15:01:01 crc kubenswrapper[4778]: I0312 15:01:01.749431 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555461-lmqk9" event={"ID":"ebdf3274-70cb-4083-bf12-5d1038a9b7ba","Type":"ContainerStarted","Data":"01dbc6ac7066fa5ead67e636626486ffc63409dac8a4cd6a20c003a2abfad4ff"} Mar 12 15:01:01 crc kubenswrapper[4778]: I0312 15:01:01.749741 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555461-lmqk9" event={"ID":"ebdf3274-70cb-4083-bf12-5d1038a9b7ba","Type":"ContainerStarted","Data":"f482217845d50b2884bd5bd48ad61af1e7768b9ee05ce69facabd403638440d6"} Mar 12 15:01:04 crc kubenswrapper[4778]: E0312 15:01:04.038076 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebdf3274_70cb_4083_bf12_5d1038a9b7ba.slice/crio-conmon-01dbc6ac7066fa5ead67e636626486ffc63409dac8a4cd6a20c003a2abfad4ff.scope\": RecentStats: unable to find data in memory cache]" Mar 12 15:01:04 crc kubenswrapper[4778]: I0312 15:01:04.800276 4778 generic.go:334] "Generic (PLEG): container finished" podID="ebdf3274-70cb-4083-bf12-5d1038a9b7ba" containerID="01dbc6ac7066fa5ead67e636626486ffc63409dac8a4cd6a20c003a2abfad4ff" exitCode=0 Mar 12 15:01:04 crc kubenswrapper[4778]: I0312 15:01:04.800423 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555461-lmqk9" event={"ID":"ebdf3274-70cb-4083-bf12-5d1038a9b7ba","Type":"ContainerDied","Data":"01dbc6ac7066fa5ead67e636626486ffc63409dac8a4cd6a20c003a2abfad4ff"} Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.206905 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.305099 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-config-data\") pod \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.305232 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-combined-ca-bundle\") pod \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.305270 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lls6t\" (UniqueName: \"kubernetes.io/projected/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-kube-api-access-lls6t\") pod \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.305318 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-fernet-keys\") pod \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\" (UID: \"ebdf3274-70cb-4083-bf12-5d1038a9b7ba\") " Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.310260 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-kube-api-access-lls6t" (OuterVolumeSpecName: "kube-api-access-lls6t") pod "ebdf3274-70cb-4083-bf12-5d1038a9b7ba" (UID: "ebdf3274-70cb-4083-bf12-5d1038a9b7ba"). InnerVolumeSpecName "kube-api-access-lls6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.311820 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ebdf3274-70cb-4083-bf12-5d1038a9b7ba" (UID: "ebdf3274-70cb-4083-bf12-5d1038a9b7ba"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.345159 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebdf3274-70cb-4083-bf12-5d1038a9b7ba" (UID: "ebdf3274-70cb-4083-bf12-5d1038a9b7ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.366424 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-config-data" (OuterVolumeSpecName: "config-data") pod "ebdf3274-70cb-4083-bf12-5d1038a9b7ba" (UID: "ebdf3274-70cb-4083-bf12-5d1038a9b7ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.407718 4778 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.407757 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.407769 4778 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.407786 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lls6t\" (UniqueName: \"kubernetes.io/projected/ebdf3274-70cb-4083-bf12-5d1038a9b7ba-kube-api-access-lls6t\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.822519 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29555461-lmqk9" event={"ID":"ebdf3274-70cb-4083-bf12-5d1038a9b7ba","Type":"ContainerDied","Data":"f482217845d50b2884bd5bd48ad61af1e7768b9ee05ce69facabd403638440d6"} Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.822888 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f482217845d50b2884bd5bd48ad61af1e7768b9ee05ce69facabd403638440d6" Mar 12 15:01:06 crc kubenswrapper[4778]: I0312 15:01:06.822562 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29555461-lmqk9" Mar 12 15:01:08 crc kubenswrapper[4778]: I0312 15:01:08.253179 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:01:08 crc kubenswrapper[4778]: E0312 15:01:08.253533 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:01:19 crc kubenswrapper[4778]: I0312 15:01:19.255047 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:01:19 crc kubenswrapper[4778]: E0312 15:01:19.258208 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:01:31 crc kubenswrapper[4778]: I0312 15:01:31.253781 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:01:31 crc kubenswrapper[4778]: E0312 15:01:31.254638 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:01:32 crc kubenswrapper[4778]: I0312 15:01:32.125150 4778 generic.go:334] "Generic (PLEG): container finished" podID="74897d0a-ca7b-4589-bd4c-75910c2d491c" containerID="04824fe8df9ecfce713c8136bfb0516b3d49f4264b49ad91474ebd09ae740d91" exitCode=0 Mar 12 15:01:32 crc kubenswrapper[4778]: I0312 15:01:32.125242 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"74897d0a-ca7b-4589-bd4c-75910c2d491c","Type":"ContainerDied","Data":"04824fe8df9ecfce713c8136bfb0516b3d49f4264b49ad91474ebd09ae740d91"} Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.864755 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.952980 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config-secret\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953091 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ssdw\" (UniqueName: \"kubernetes.io/projected/74897d0a-ca7b-4589-bd4c-75910c2d491c-kube-api-access-4ssdw\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953132 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ssh-key\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953172 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-workdir\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953235 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-config-data\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953293 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953419 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953529 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-temporary\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.953590 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ca-certs\") pod \"74897d0a-ca7b-4589-bd4c-75910c2d491c\" (UID: \"74897d0a-ca7b-4589-bd4c-75910c2d491c\") " Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.955227 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-config-data" (OuterVolumeSpecName: "config-data") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.955427 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.959093 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.959460 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74897d0a-ca7b-4589-bd4c-75910c2d491c-kube-api-access-4ssdw" (OuterVolumeSpecName: "kube-api-access-4ssdw") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "kube-api-access-4ssdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.964140 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.983963 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.994030 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:01:33 crc kubenswrapper[4778]: I0312 15:01:33.998360 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.035023 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "74897d0a-ca7b-4589-bd4c-75910c2d491c" (UID: "74897d0a-ca7b-4589-bd4c-75910c2d491c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056631 4778 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056676 4778 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056692 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056706 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ssdw\" (UniqueName: \"kubernetes.io/projected/74897d0a-ca7b-4589-bd4c-75910c2d491c-kube-api-access-4ssdw\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056718 4778 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74897d0a-ca7b-4589-bd4c-75910c2d491c-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056731 4778 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/74897d0a-ca7b-4589-bd4c-75910c2d491c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056745 4778 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056758 4778 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74897d0a-ca7b-4589-bd4c-75910c2d491c-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.056800 4778 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.086885 4778 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.155371 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"74897d0a-ca7b-4589-bd4c-75910c2d491c","Type":"ContainerDied","Data":"454ca901956127a4048551d166d33c00269e2d8a18f508b4b327654529c385c0"} Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.155423 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.155427 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="454ca901956127a4048551d166d33c00269e2d8a18f508b4b327654529c385c0" Mar 12 15:01:34 crc kubenswrapper[4778]: I0312 15:01:34.158288 4778 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.768727 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 12 15:01:39 crc kubenswrapper[4778]: E0312 15:01:39.769936 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74897d0a-ca7b-4589-bd4c-75910c2d491c" containerName="tempest-tests-tempest-tests-runner" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.769955 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="74897d0a-ca7b-4589-bd4c-75910c2d491c" containerName="tempest-tests-tempest-tests-runner" Mar 12 15:01:39 crc kubenswrapper[4778]: E0312 15:01:39.769977 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdf3274-70cb-4083-bf12-5d1038a9b7ba" containerName="keystone-cron" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.769986 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdf3274-70cb-4083-bf12-5d1038a9b7ba" containerName="keystone-cron" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.770234 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebdf3274-70cb-4083-bf12-5d1038a9b7ba" containerName="keystone-cron" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.770258 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="74897d0a-ca7b-4589-bd4c-75910c2d491c" containerName="tempest-tests-tempest-tests-runner" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.771026 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.773970 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-s8dkq" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.782698 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.882264 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jtd4\" (UniqueName: \"kubernetes.io/projected/82246f69-2112-44e9-a783-a4a5926188b4-kube-api-access-2jtd4\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.882331 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.985476 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jtd4\" (UniqueName: \"kubernetes.io/projected/82246f69-2112-44e9-a783-a4a5926188b4-kube-api-access-2jtd4\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.985607 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:39 crc kubenswrapper[4778]: I0312 15:01:39.986313 4778 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:40 crc kubenswrapper[4778]: I0312 15:01:40.012311 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jtd4\" (UniqueName: \"kubernetes.io/projected/82246f69-2112-44e9-a783-a4a5926188b4-kube-api-access-2jtd4\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:40 crc kubenswrapper[4778]: I0312 15:01:40.036125 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82246f69-2112-44e9-a783-a4a5926188b4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:40 crc kubenswrapper[4778]: I0312 15:01:40.115741 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 12 15:01:40 crc kubenswrapper[4778]: I0312 15:01:40.607932 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 12 15:01:41 crc kubenswrapper[4778]: I0312 15:01:41.234577 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"82246f69-2112-44e9-a783-a4a5926188b4","Type":"ContainerStarted","Data":"86122024dcb612e716c8156c8914ce0d795dee3075f7fd0fe85a17b803420332"} Mar 12 15:01:42 crc kubenswrapper[4778]: I0312 15:01:42.243985 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"82246f69-2112-44e9-a783-a4a5926188b4","Type":"ContainerStarted","Data":"6a1da5a5609d7d84e85854d47472704f89202cab80268f35990e8e18b239063f"} Mar 12 15:01:42 crc kubenswrapper[4778]: I0312 15:01:42.270343 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.9905026289999999 podStartE2EDuration="3.270316627s" podCreationTimestamp="2026-03-12 15:01:39 +0000 UTC" firstStartedPulling="2026-03-12 15:01:40.607062704 +0000 UTC m=+6719.055758100" lastFinishedPulling="2026-03-12 15:01:41.886876692 +0000 UTC m=+6720.335572098" observedRunningTime="2026-03-12 15:01:42.264820341 +0000 UTC m=+6720.713515737" watchObservedRunningTime="2026-03-12 15:01:42.270316627 +0000 UTC m=+6720.719012063" Mar 12 15:01:44 crc kubenswrapper[4778]: I0312 15:01:44.254308 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:01:44 crc kubenswrapper[4778]: E0312 15:01:44.255228 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:01:57 crc kubenswrapper[4778]: I0312 15:01:57.255528 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:01:57 crc kubenswrapper[4778]: E0312 15:01:57.256339 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.162589 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555462-dvtsm"] Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.165693 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.169898 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.170564 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.170882 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.191259 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555462-dvtsm"] Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.243969 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxkr6\" (UniqueName: \"kubernetes.io/projected/cbb23378-6e3e-4c63-919d-47ce1d17dd7b-kube-api-access-nxkr6\") pod \"auto-csr-approver-29555462-dvtsm\" (UID: \"cbb23378-6e3e-4c63-919d-47ce1d17dd7b\") " pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.346048 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxkr6\" (UniqueName: \"kubernetes.io/projected/cbb23378-6e3e-4c63-919d-47ce1d17dd7b-kube-api-access-nxkr6\") pod \"auto-csr-approver-29555462-dvtsm\" (UID: \"cbb23378-6e3e-4c63-919d-47ce1d17dd7b\") " pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.366292 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxkr6\" (UniqueName: \"kubernetes.io/projected/cbb23378-6e3e-4c63-919d-47ce1d17dd7b-kube-api-access-nxkr6\") pod \"auto-csr-approver-29555462-dvtsm\" (UID: \"cbb23378-6e3e-4c63-919d-47ce1d17dd7b\") " pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.500270 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:00 crc kubenswrapper[4778]: I0312 15:02:00.976633 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555462-dvtsm"] Mar 12 15:02:01 crc kubenswrapper[4778]: I0312 15:02:01.438876 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" event={"ID":"cbb23378-6e3e-4c63-919d-47ce1d17dd7b","Type":"ContainerStarted","Data":"deaecc829dedc237b70dc46e8c5b40e55230025755e171dfd88c21e6975391c5"} Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.402456 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rkpvq/must-gather-6d9ls"] Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.405003 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.406915 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rkpvq"/"default-dockercfg-fscgd" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.407345 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rkpvq"/"kube-root-ca.crt" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.410601 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rkpvq"/"openshift-service-ca.crt" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.415034 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rkpvq/must-gather-6d9ls"] Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.459440 4778 generic.go:334] "Generic (PLEG): container finished" podID="cbb23378-6e3e-4c63-919d-47ce1d17dd7b" containerID="bc1c69d732ac8380ce4ad84b76897a91373ec3edde2343f57d27f4105f4594eb" exitCode=0 Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.459481 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" event={"ID":"cbb23378-6e3e-4c63-919d-47ce1d17dd7b","Type":"ContainerDied","Data":"bc1c69d732ac8380ce4ad84b76897a91373ec3edde2343f57d27f4105f4594eb"} Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.510323 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twzlp\" (UniqueName: \"kubernetes.io/projected/dd2baa0b-6680-41af-8231-e30368cb0090-kube-api-access-twzlp\") pod \"must-gather-6d9ls\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.510513 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dd2baa0b-6680-41af-8231-e30368cb0090-must-gather-output\") pod \"must-gather-6d9ls\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.612263 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dd2baa0b-6680-41af-8231-e30368cb0090-must-gather-output\") pod \"must-gather-6d9ls\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.612342 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twzlp\" (UniqueName: \"kubernetes.io/projected/dd2baa0b-6680-41af-8231-e30368cb0090-kube-api-access-twzlp\") pod \"must-gather-6d9ls\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.612759 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dd2baa0b-6680-41af-8231-e30368cb0090-must-gather-output\") pod \"must-gather-6d9ls\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.629707 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twzlp\" (UniqueName: \"kubernetes.io/projected/dd2baa0b-6680-41af-8231-e30368cb0090-kube-api-access-twzlp\") pod \"must-gather-6d9ls\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:03 crc kubenswrapper[4778]: I0312 15:02:03.725624 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:02:04 crc kubenswrapper[4778]: I0312 15:02:04.232897 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rkpvq/must-gather-6d9ls"] Mar 12 15:02:04 crc kubenswrapper[4778]: I0312 15:02:04.470351 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" event={"ID":"dd2baa0b-6680-41af-8231-e30368cb0090","Type":"ContainerStarted","Data":"49571a492d3a83d3d165f3c9920027fe7d0f5cf5c45cb08c90ad79451a4a973a"} Mar 12 15:02:04 crc kubenswrapper[4778]: I0312 15:02:04.842144 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:04 crc kubenswrapper[4778]: I0312 15:02:04.938020 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxkr6\" (UniqueName: \"kubernetes.io/projected/cbb23378-6e3e-4c63-919d-47ce1d17dd7b-kube-api-access-nxkr6\") pod \"cbb23378-6e3e-4c63-919d-47ce1d17dd7b\" (UID: \"cbb23378-6e3e-4c63-919d-47ce1d17dd7b\") " Mar 12 15:02:04 crc kubenswrapper[4778]: I0312 15:02:04.945196 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb23378-6e3e-4c63-919d-47ce1d17dd7b-kube-api-access-nxkr6" (OuterVolumeSpecName: "kube-api-access-nxkr6") pod "cbb23378-6e3e-4c63-919d-47ce1d17dd7b" (UID: "cbb23378-6e3e-4c63-919d-47ce1d17dd7b"). InnerVolumeSpecName "kube-api-access-nxkr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:02:05 crc kubenswrapper[4778]: I0312 15:02:05.040689 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxkr6\" (UniqueName: \"kubernetes.io/projected/cbb23378-6e3e-4c63-919d-47ce1d17dd7b-kube-api-access-nxkr6\") on node \"crc\" DevicePath \"\"" Mar 12 15:02:05 crc kubenswrapper[4778]: I0312 15:02:05.512636 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" event={"ID":"cbb23378-6e3e-4c63-919d-47ce1d17dd7b","Type":"ContainerDied","Data":"deaecc829dedc237b70dc46e8c5b40e55230025755e171dfd88c21e6975391c5"} Mar 12 15:02:05 crc kubenswrapper[4778]: I0312 15:02:05.512714 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deaecc829dedc237b70dc46e8c5b40e55230025755e171dfd88c21e6975391c5" Mar 12 15:02:05 crc kubenswrapper[4778]: I0312 15:02:05.512717 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555462-dvtsm" Mar 12 15:02:05 crc kubenswrapper[4778]: E0312 15:02:05.724908 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb23378_6e3e_4c63_919d_47ce1d17dd7b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb23378_6e3e_4c63_919d_47ce1d17dd7b.slice/crio-deaecc829dedc237b70dc46e8c5b40e55230025755e171dfd88c21e6975391c5\": RecentStats: unable to find data in memory cache]" Mar 12 15:02:05 crc kubenswrapper[4778]: I0312 15:02:05.917455 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555456-vtfp6"] Mar 12 15:02:05 crc kubenswrapper[4778]: I0312 15:02:05.928448 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555456-vtfp6"] Mar 12 15:02:06 crc kubenswrapper[4778]: I0312 15:02:06.265615 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0add53-7611-4f91-bf0b-cf5fea5bb9d7" path="/var/lib/kubelet/pods/fc0add53-7611-4f91-bf0b-cf5fea5bb9d7/volumes" Mar 12 15:02:08 crc kubenswrapper[4778]: I0312 15:02:08.254713 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:02:10 crc kubenswrapper[4778]: I0312 15:02:10.561627 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" event={"ID":"dd2baa0b-6680-41af-8231-e30368cb0090","Type":"ContainerStarted","Data":"099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb"} Mar 12 15:02:10 crc kubenswrapper[4778]: I0312 15:02:10.563402 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"009e612c3693545ba4a1988aa00993d05612427ec6eb485b08b455b35968f1ab"} Mar 12 15:02:11 crc kubenswrapper[4778]: I0312 15:02:11.574428 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" event={"ID":"dd2baa0b-6680-41af-8231-e30368cb0090","Type":"ContainerStarted","Data":"d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b"} Mar 12 15:02:11 crc kubenswrapper[4778]: I0312 15:02:11.595226 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" podStartSLOduration=2.649630095 podStartE2EDuration="8.595205017s" podCreationTimestamp="2026-03-12 15:02:03 +0000 UTC" firstStartedPulling="2026-03-12 15:02:04.234699723 +0000 UTC m=+6742.683395119" lastFinishedPulling="2026-03-12 15:02:10.180274625 +0000 UTC m=+6748.628970041" observedRunningTime="2026-03-12 15:02:11.587812166 +0000 UTC m=+6750.036507562" watchObservedRunningTime="2026-03-12 15:02:11.595205017 +0000 UTC m=+6750.043900413" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.272817 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-2n5vv"] Mar 12 15:02:15 crc kubenswrapper[4778]: E0312 15:02:15.273545 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb23378-6e3e-4c63-919d-47ce1d17dd7b" containerName="oc" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.273560 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb23378-6e3e-4c63-919d-47ce1d17dd7b" containerName="oc" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.273769 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb23378-6e3e-4c63-919d-47ce1d17dd7b" containerName="oc" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.274357 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.370794 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-host\") pod \"crc-debug-2n5vv\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.370883 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9m8w\" (UniqueName: \"kubernetes.io/projected/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-kube-api-access-d9m8w\") pod \"crc-debug-2n5vv\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.473365 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9m8w\" (UniqueName: \"kubernetes.io/projected/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-kube-api-access-d9m8w\") pod \"crc-debug-2n5vv\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.473726 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-host\") pod \"crc-debug-2n5vv\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.473827 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-host\") pod \"crc-debug-2n5vv\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.497196 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9m8w\" (UniqueName: \"kubernetes.io/projected/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-kube-api-access-d9m8w\") pod \"crc-debug-2n5vv\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:15 crc kubenswrapper[4778]: I0312 15:02:15.594967 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:02:16 crc kubenswrapper[4778]: I0312 15:02:16.623680 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" event={"ID":"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657","Type":"ContainerStarted","Data":"94f570da92f9fd8977fa9f3cbc8d1dfd2eab1360e49ff3877e319f0a0fea4cfa"} Mar 12 15:02:24 crc kubenswrapper[4778]: I0312 15:02:24.705914 4778 scope.go:117] "RemoveContainer" containerID="d7d76c5b2f5b6d4767497e4e99746de9373b74f615023933a47cba956a1bacb0" Mar 12 15:02:26 crc kubenswrapper[4778]: I0312 15:02:26.714522 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" event={"ID":"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657","Type":"ContainerStarted","Data":"32e69f17a15da926e453ec4388e2482d274516709eb37f6124496feae6a6509f"} Mar 12 15:02:26 crc kubenswrapper[4778]: I0312 15:02:26.728553 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" podStartSLOduration=1.746772557 podStartE2EDuration="11.728538109s" podCreationTimestamp="2026-03-12 15:02:15 +0000 UTC" firstStartedPulling="2026-03-12 15:02:15.657244171 +0000 UTC m=+6754.105939587" lastFinishedPulling="2026-03-12 15:02:25.639009743 +0000 UTC m=+6764.087705139" observedRunningTime="2026-03-12 15:02:26.728504578 +0000 UTC m=+6765.177199974" watchObservedRunningTime="2026-03-12 15:02:26.728538109 +0000 UTC m=+6765.177233505" Mar 12 15:02:42 crc kubenswrapper[4778]: I0312 15:02:42.847810 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h8jzw"] Mar 12 15:02:42 crc kubenswrapper[4778]: I0312 15:02:42.850714 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:42 crc kubenswrapper[4778]: I0312 15:02:42.868220 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8jzw"] Mar 12 15:02:42 crc kubenswrapper[4778]: I0312 15:02:42.963917 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-catalog-content\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:42 crc kubenswrapper[4778]: I0312 15:02:42.964414 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24lfq\" (UniqueName: \"kubernetes.io/projected/9c3f77a6-b73d-4572-9e3c-57622161ebab-kube-api-access-24lfq\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:42 crc kubenswrapper[4778]: I0312 15:02:42.964644 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-utilities\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.066256 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24lfq\" (UniqueName: \"kubernetes.io/projected/9c3f77a6-b73d-4572-9e3c-57622161ebab-kube-api-access-24lfq\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.066355 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-utilities\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.066409 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-catalog-content\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.067046 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-catalog-content\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.067073 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-utilities\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.096584 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24lfq\" (UniqueName: \"kubernetes.io/projected/9c3f77a6-b73d-4572-9e3c-57622161ebab-kube-api-access-24lfq\") pod \"community-operators-h8jzw\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.173445 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.735585 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8jzw"] Mar 12 15:02:43 crc kubenswrapper[4778]: I0312 15:02:43.883273 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerStarted","Data":"bc58286ff5647c69df9b0e6066c0da0456949211defbad08ed82bfddbfdcb9d5"} Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.652018 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wfdzw"] Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.655949 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.665557 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfdzw"] Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.700585 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-catalog-content\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.700680 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnxq6\" (UniqueName: \"kubernetes.io/projected/d8e947a7-cded-4a65-9b13-96116f14554a-kube-api-access-mnxq6\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.700715 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-utilities\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.802156 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-catalog-content\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.802248 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnxq6\" (UniqueName: \"kubernetes.io/projected/d8e947a7-cded-4a65-9b13-96116f14554a-kube-api-access-mnxq6\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.802274 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-utilities\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.802773 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-utilities\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.802998 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-catalog-content\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.826850 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnxq6\" (UniqueName: \"kubernetes.io/projected/d8e947a7-cded-4a65-9b13-96116f14554a-kube-api-access-mnxq6\") pod \"redhat-marketplace-wfdzw\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.893831 4778 generic.go:334] "Generic (PLEG): container finished" podID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerID="e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f" exitCode=0 Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.893869 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerDied","Data":"e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f"} Mar 12 15:02:44 crc kubenswrapper[4778]: I0312 15:02:44.983755 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:45 crc kubenswrapper[4778]: I0312 15:02:45.432914 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfdzw"] Mar 12 15:02:45 crc kubenswrapper[4778]: I0312 15:02:45.905546 4778 generic.go:334] "Generic (PLEG): container finished" podID="d8e947a7-cded-4a65-9b13-96116f14554a" containerID="ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517" exitCode=0 Mar 12 15:02:45 crc kubenswrapper[4778]: I0312 15:02:45.905703 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerDied","Data":"ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517"} Mar 12 15:02:45 crc kubenswrapper[4778]: I0312 15:02:45.906088 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerStarted","Data":"fc06856e4d46053cfe775ea619396f83fa56c6a09a352c79ce3aaf57d3b1e242"} Mar 12 15:02:45 crc kubenswrapper[4778]: I0312 15:02:45.912255 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerStarted","Data":"8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91"} Mar 12 15:02:46 crc kubenswrapper[4778]: E0312 15:02:46.801629 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c3f77a6_b73d_4572_9e3c_57622161ebab.slice/crio-conmon-8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91.scope\": RecentStats: unable to find data in memory cache]" Mar 12 15:02:46 crc kubenswrapper[4778]: I0312 15:02:46.923350 4778 generic.go:334] "Generic (PLEG): container finished" podID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerID="8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91" exitCode=0 Mar 12 15:02:46 crc kubenswrapper[4778]: I0312 15:02:46.923477 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerDied","Data":"8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91"} Mar 12 15:02:47 crc kubenswrapper[4778]: I0312 15:02:47.935239 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerStarted","Data":"825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b"} Mar 12 15:02:47 crc kubenswrapper[4778]: I0312 15:02:47.938754 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerStarted","Data":"e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822"} Mar 12 15:02:47 crc kubenswrapper[4778]: I0312 15:02:47.990396 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h8jzw" podStartSLOduration=3.453537838 podStartE2EDuration="5.990371287s" podCreationTimestamp="2026-03-12 15:02:42 +0000 UTC" firstStartedPulling="2026-03-12 15:02:44.895915289 +0000 UTC m=+6783.344610685" lastFinishedPulling="2026-03-12 15:02:47.432748738 +0000 UTC m=+6785.881444134" observedRunningTime="2026-03-12 15:02:47.989200033 +0000 UTC m=+6786.437895429" watchObservedRunningTime="2026-03-12 15:02:47.990371287 +0000 UTC m=+6786.439066683" Mar 12 15:02:48 crc kubenswrapper[4778]: I0312 15:02:48.947927 4778 generic.go:334] "Generic (PLEG): container finished" podID="d8e947a7-cded-4a65-9b13-96116f14554a" containerID="825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b" exitCode=0 Mar 12 15:02:48 crc kubenswrapper[4778]: I0312 15:02:48.949361 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerDied","Data":"825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b"} Mar 12 15:02:49 crc kubenswrapper[4778]: I0312 15:02:49.958953 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerStarted","Data":"d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc"} Mar 12 15:02:49 crc kubenswrapper[4778]: I0312 15:02:49.986656 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wfdzw" podStartSLOduration=2.4527912179999998 podStartE2EDuration="5.98663979s" podCreationTimestamp="2026-03-12 15:02:44 +0000 UTC" firstStartedPulling="2026-03-12 15:02:45.907547591 +0000 UTC m=+6784.356242987" lastFinishedPulling="2026-03-12 15:02:49.441396163 +0000 UTC m=+6787.890091559" observedRunningTime="2026-03-12 15:02:49.97714872 +0000 UTC m=+6788.425844116" watchObservedRunningTime="2026-03-12 15:02:49.98663979 +0000 UTC m=+6788.435335186" Mar 12 15:02:53 crc kubenswrapper[4778]: I0312 15:02:53.176599 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:53 crc kubenswrapper[4778]: I0312 15:02:53.177147 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:02:54 crc kubenswrapper[4778]: I0312 15:02:54.236673 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h8jzw" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="registry-server" probeResult="failure" output=< Mar 12 15:02:54 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 15:02:54 crc kubenswrapper[4778]: > Mar 12 15:02:54 crc kubenswrapper[4778]: I0312 15:02:54.984538 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:54 crc kubenswrapper[4778]: I0312 15:02:54.984608 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:02:56 crc kubenswrapper[4778]: I0312 15:02:56.038779 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-wfdzw" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="registry-server" probeResult="failure" output=< Mar 12 15:02:56 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 15:02:56 crc kubenswrapper[4778]: > Mar 12 15:03:03 crc kubenswrapper[4778]: I0312 15:03:03.225897 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:03:03 crc kubenswrapper[4778]: I0312 15:03:03.293496 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:03:03 crc kubenswrapper[4778]: I0312 15:03:03.462443 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h8jzw"] Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.039526 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.083703 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.093467 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h8jzw" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="registry-server" containerID="cri-o://e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822" gracePeriod=2 Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.616506 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.737410 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-catalog-content\") pod \"9c3f77a6-b73d-4572-9e3c-57622161ebab\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.737571 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24lfq\" (UniqueName: \"kubernetes.io/projected/9c3f77a6-b73d-4572-9e3c-57622161ebab-kube-api-access-24lfq\") pod \"9c3f77a6-b73d-4572-9e3c-57622161ebab\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.737703 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-utilities\") pod \"9c3f77a6-b73d-4572-9e3c-57622161ebab\" (UID: \"9c3f77a6-b73d-4572-9e3c-57622161ebab\") " Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.738456 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-utilities" (OuterVolumeSpecName: "utilities") pod "9c3f77a6-b73d-4572-9e3c-57622161ebab" (UID: "9c3f77a6-b73d-4572-9e3c-57622161ebab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.744679 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c3f77a6-b73d-4572-9e3c-57622161ebab-kube-api-access-24lfq" (OuterVolumeSpecName: "kube-api-access-24lfq") pod "9c3f77a6-b73d-4572-9e3c-57622161ebab" (UID: "9c3f77a6-b73d-4572-9e3c-57622161ebab"). InnerVolumeSpecName "kube-api-access-24lfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.797911 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c3f77a6-b73d-4572-9e3c-57622161ebab" (UID: "9c3f77a6-b73d-4572-9e3c-57622161ebab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.840399 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24lfq\" (UniqueName: \"kubernetes.io/projected/9c3f77a6-b73d-4572-9e3c-57622161ebab-kube-api-access-24lfq\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.840436 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:05 crc kubenswrapper[4778]: I0312 15:03:05.840447 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3f77a6-b73d-4572-9e3c-57622161ebab-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.060991 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfdzw"] Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.102106 4778 generic.go:334] "Generic (PLEG): container finished" podID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerID="e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822" exitCode=0 Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.102290 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wfdzw" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="registry-server" containerID="cri-o://d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc" gracePeriod=2 Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.102587 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8jzw" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.107258 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerDied","Data":"e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822"} Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.107304 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8jzw" event={"ID":"9c3f77a6-b73d-4572-9e3c-57622161ebab","Type":"ContainerDied","Data":"bc58286ff5647c69df9b0e6066c0da0456949211defbad08ed82bfddbfdcb9d5"} Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.107326 4778 scope.go:117] "RemoveContainer" containerID="e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.143646 4778 scope.go:117] "RemoveContainer" containerID="8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.159062 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h8jzw"] Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.167423 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h8jzw"] Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.171437 4778 scope.go:117] "RemoveContainer" containerID="e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.268470 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" path="/var/lib/kubelet/pods/9c3f77a6-b73d-4572-9e3c-57622161ebab/volumes" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.326853 4778 scope.go:117] "RemoveContainer" containerID="e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822" Mar 12 15:03:06 crc kubenswrapper[4778]: E0312 15:03:06.327568 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822\": container with ID starting with e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822 not found: ID does not exist" containerID="e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.327598 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822"} err="failed to get container status \"e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822\": rpc error: code = NotFound desc = could not find container \"e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822\": container with ID starting with e2eaa7378d99f8573e564b6e7f6cad396b4f6de5d0a6f8d2464a5ad678335822 not found: ID does not exist" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.327618 4778 scope.go:117] "RemoveContainer" containerID="8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91" Mar 12 15:03:06 crc kubenswrapper[4778]: E0312 15:03:06.328588 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91\": container with ID starting with 8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91 not found: ID does not exist" containerID="8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.328641 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91"} err="failed to get container status \"8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91\": rpc error: code = NotFound desc = could not find container \"8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91\": container with ID starting with 8299d570ee7a131ad1e48e458fb394fad06615e92119ee1cd3fcb7838dce6a91 not found: ID does not exist" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.328676 4778 scope.go:117] "RemoveContainer" containerID="e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f" Mar 12 15:03:06 crc kubenswrapper[4778]: E0312 15:03:06.328924 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f\": container with ID starting with e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f not found: ID does not exist" containerID="e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.328953 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f"} err="failed to get container status \"e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f\": rpc error: code = NotFound desc = could not find container \"e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f\": container with ID starting with e805e212c2a8d910e4812597491e17968ae4969f60f7e1630b50cf7c475b216f not found: ID does not exist" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.543834 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.658957 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-catalog-content\") pod \"d8e947a7-cded-4a65-9b13-96116f14554a\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.659237 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-utilities\") pod \"d8e947a7-cded-4a65-9b13-96116f14554a\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.659290 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnxq6\" (UniqueName: \"kubernetes.io/projected/d8e947a7-cded-4a65-9b13-96116f14554a-kube-api-access-mnxq6\") pod \"d8e947a7-cded-4a65-9b13-96116f14554a\" (UID: \"d8e947a7-cded-4a65-9b13-96116f14554a\") " Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.661920 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-utilities" (OuterVolumeSpecName: "utilities") pod "d8e947a7-cded-4a65-9b13-96116f14554a" (UID: "d8e947a7-cded-4a65-9b13-96116f14554a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.665229 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e947a7-cded-4a65-9b13-96116f14554a-kube-api-access-mnxq6" (OuterVolumeSpecName: "kube-api-access-mnxq6") pod "d8e947a7-cded-4a65-9b13-96116f14554a" (UID: "d8e947a7-cded-4a65-9b13-96116f14554a"). InnerVolumeSpecName "kube-api-access-mnxq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.694329 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8e947a7-cded-4a65-9b13-96116f14554a" (UID: "d8e947a7-cded-4a65-9b13-96116f14554a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.761084 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.761124 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e947a7-cded-4a65-9b13-96116f14554a-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:06 crc kubenswrapper[4778]: I0312 15:03:06.761135 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnxq6\" (UniqueName: \"kubernetes.io/projected/d8e947a7-cded-4a65-9b13-96116f14554a-kube-api-access-mnxq6\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.114965 4778 generic.go:334] "Generic (PLEG): container finished" podID="d8e947a7-cded-4a65-9b13-96116f14554a" containerID="d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc" exitCode=0 Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.115136 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerDied","Data":"d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc"} Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.115198 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfdzw" event={"ID":"d8e947a7-cded-4a65-9b13-96116f14554a","Type":"ContainerDied","Data":"fc06856e4d46053cfe775ea619396f83fa56c6a09a352c79ce3aaf57d3b1e242"} Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.115216 4778 scope.go:117] "RemoveContainer" containerID="d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.116809 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfdzw" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.146345 4778 scope.go:117] "RemoveContainer" containerID="825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.172531 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfdzw"] Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.182922 4778 scope.go:117] "RemoveContainer" containerID="ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.183423 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfdzw"] Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.211753 4778 scope.go:117] "RemoveContainer" containerID="d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc" Mar 12 15:03:07 crc kubenswrapper[4778]: E0312 15:03:07.212365 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc\": container with ID starting with d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc not found: ID does not exist" containerID="d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.212423 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc"} err="failed to get container status \"d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc\": rpc error: code = NotFound desc = could not find container \"d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc\": container with ID starting with d8872ce63bbd32f8c586ac1d8dd82e4953cef5369d5780daadde079262b016fc not found: ID does not exist" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.212455 4778 scope.go:117] "RemoveContainer" containerID="825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b" Mar 12 15:03:07 crc kubenswrapper[4778]: E0312 15:03:07.215068 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b\": container with ID starting with 825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b not found: ID does not exist" containerID="825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.215130 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b"} err="failed to get container status \"825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b\": rpc error: code = NotFound desc = could not find container \"825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b\": container with ID starting with 825a3ee224fa4e45b5da52433ce6b8e182cae07f2f9fc2009e00554dfb8ffd2b not found: ID does not exist" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.215164 4778 scope.go:117] "RemoveContainer" containerID="ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517" Mar 12 15:03:07 crc kubenswrapper[4778]: E0312 15:03:07.215696 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517\": container with ID starting with ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517 not found: ID does not exist" containerID="ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517" Mar 12 15:03:07 crc kubenswrapper[4778]: I0312 15:03:07.215731 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517"} err="failed to get container status \"ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517\": rpc error: code = NotFound desc = could not find container \"ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517\": container with ID starting with ca2b00af1cbd99c0d40d91469371bc813d95cbd3211d2c77ddb4bffd816ae517 not found: ID does not exist" Mar 12 15:03:07 crc kubenswrapper[4778]: E0312 15:03:07.344811 4778 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8e947a7_cded_4a65_9b13_96116f14554a.slice/crio-fc06856e4d46053cfe775ea619396f83fa56c6a09a352c79ce3aaf57d3b1e242\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8e947a7_cded_4a65_9b13_96116f14554a.slice\": RecentStats: unable to find data in memory cache]" Mar 12 15:03:08 crc kubenswrapper[4778]: I0312 15:03:08.267677 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" path="/var/lib/kubelet/pods/d8e947a7-cded-4a65-9b13-96116f14554a/volumes" Mar 12 15:03:11 crc kubenswrapper[4778]: I0312 15:03:11.160341 4778 generic.go:334] "Generic (PLEG): container finished" podID="25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" containerID="32e69f17a15da926e453ec4388e2482d274516709eb37f6124496feae6a6509f" exitCode=0 Mar 12 15:03:11 crc kubenswrapper[4778]: I0312 15:03:11.160434 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" event={"ID":"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657","Type":"ContainerDied","Data":"32e69f17a15da926e453ec4388e2482d274516709eb37f6124496feae6a6509f"} Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.285705 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.317092 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-2n5vv"] Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.329011 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-2n5vv"] Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.478778 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-host\") pod \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.478855 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9m8w\" (UniqueName: \"kubernetes.io/projected/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-kube-api-access-d9m8w\") pod \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\" (UID: \"25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657\") " Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.479337 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-host" (OuterVolumeSpecName: "host") pod "25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" (UID: "25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.479702 4778 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-host\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.485978 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-kube-api-access-d9m8w" (OuterVolumeSpecName: "kube-api-access-d9m8w") pod "25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" (UID: "25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657"). InnerVolumeSpecName "kube-api-access-d9m8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:03:12 crc kubenswrapper[4778]: I0312 15:03:12.581995 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9m8w\" (UniqueName: \"kubernetes.io/projected/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657-kube-api-access-d9m8w\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.178668 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94f570da92f9fd8977fa9f3cbc8d1dfd2eab1360e49ff3877e319f0a0fea4cfa" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.178730 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-2n5vv" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.506783 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-skmxt"] Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507289 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="registry-server" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507306 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="registry-server" Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507325 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="extract-utilities" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507333 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="extract-utilities" Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507398 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="extract-utilities" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507409 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="extract-utilities" Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507450 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="registry-server" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507459 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="registry-server" Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507475 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="extract-content" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507482 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="extract-content" Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507501 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" containerName="container-00" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507509 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" containerName="container-00" Mar 12 15:03:13 crc kubenswrapper[4778]: E0312 15:03:13.507520 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="extract-content" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507528 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="extract-content" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507770 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3f77a6-b73d-4572-9e3c-57622161ebab" containerName="registry-server" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507790 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" containerName="container-00" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.507809 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e947a7-cded-4a65-9b13-96116f14554a" containerName="registry-server" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.508517 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.700904 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5djj\" (UniqueName: \"kubernetes.io/projected/434eb649-8d17-4d84-977f-a1907290d0f4-kube-api-access-x5djj\") pod \"crc-debug-skmxt\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.700968 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/434eb649-8d17-4d84-977f-a1907290d0f4-host\") pod \"crc-debug-skmxt\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.803631 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5djj\" (UniqueName: \"kubernetes.io/projected/434eb649-8d17-4d84-977f-a1907290d0f4-kube-api-access-x5djj\") pod \"crc-debug-skmxt\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.803724 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/434eb649-8d17-4d84-977f-a1907290d0f4-host\") pod \"crc-debug-skmxt\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.803825 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/434eb649-8d17-4d84-977f-a1907290d0f4-host\") pod \"crc-debug-skmxt\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.823519 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5djj\" (UniqueName: \"kubernetes.io/projected/434eb649-8d17-4d84-977f-a1907290d0f4-kube-api-access-x5djj\") pod \"crc-debug-skmxt\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:13 crc kubenswrapper[4778]: I0312 15:03:13.828671 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:14 crc kubenswrapper[4778]: I0312 15:03:14.188508 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" event={"ID":"434eb649-8d17-4d84-977f-a1907290d0f4","Type":"ContainerStarted","Data":"e0b26a87b52c43c4e608ab6f59fc6362851c87a5e28dc029fccea83b1c3e5e7d"} Mar 12 15:03:14 crc kubenswrapper[4778]: I0312 15:03:14.189040 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" event={"ID":"434eb649-8d17-4d84-977f-a1907290d0f4","Type":"ContainerStarted","Data":"00e4b5ab7c733c109c6095f1277f619b9996f581fc3689633c33016f34739ddf"} Mar 12 15:03:14 crc kubenswrapper[4778]: I0312 15:03:14.216468 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" podStartSLOduration=1.216445107 podStartE2EDuration="1.216445107s" podCreationTimestamp="2026-03-12 15:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 15:03:14.203278922 +0000 UTC m=+6812.651974328" watchObservedRunningTime="2026-03-12 15:03:14.216445107 +0000 UTC m=+6812.665140503" Mar 12 15:03:14 crc kubenswrapper[4778]: I0312 15:03:14.276098 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657" path="/var/lib/kubelet/pods/25b8a8bd-e1c2-44d9-8a16-a1efc4aaf657/volumes" Mar 12 15:03:15 crc kubenswrapper[4778]: I0312 15:03:15.199291 4778 generic.go:334] "Generic (PLEG): container finished" podID="434eb649-8d17-4d84-977f-a1907290d0f4" containerID="e0b26a87b52c43c4e608ab6f59fc6362851c87a5e28dc029fccea83b1c3e5e7d" exitCode=0 Mar 12 15:03:15 crc kubenswrapper[4778]: I0312 15:03:15.199580 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" event={"ID":"434eb649-8d17-4d84-977f-a1907290d0f4","Type":"ContainerDied","Data":"e0b26a87b52c43c4e608ab6f59fc6362851c87a5e28dc029fccea83b1c3e5e7d"} Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.307434 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.356852 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-skmxt"] Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.364558 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-skmxt"] Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.451424 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/434eb649-8d17-4d84-977f-a1907290d0f4-host\") pod \"434eb649-8d17-4d84-977f-a1907290d0f4\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.451566 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5djj\" (UniqueName: \"kubernetes.io/projected/434eb649-8d17-4d84-977f-a1907290d0f4-kube-api-access-x5djj\") pod \"434eb649-8d17-4d84-977f-a1907290d0f4\" (UID: \"434eb649-8d17-4d84-977f-a1907290d0f4\") " Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.452523 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/434eb649-8d17-4d84-977f-a1907290d0f4-host" (OuterVolumeSpecName: "host") pod "434eb649-8d17-4d84-977f-a1907290d0f4" (UID: "434eb649-8d17-4d84-977f-a1907290d0f4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.461063 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434eb649-8d17-4d84-977f-a1907290d0f4-kube-api-access-x5djj" (OuterVolumeSpecName: "kube-api-access-x5djj") pod "434eb649-8d17-4d84-977f-a1907290d0f4" (UID: "434eb649-8d17-4d84-977f-a1907290d0f4"). InnerVolumeSpecName "kube-api-access-x5djj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.554012 4778 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/434eb649-8d17-4d84-977f-a1907290d0f4-host\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:16 crc kubenswrapper[4778]: I0312 15:03:16.554353 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5djj\" (UniqueName: \"kubernetes.io/projected/434eb649-8d17-4d84-977f-a1907290d0f4-kube-api-access-x5djj\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.216743 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00e4b5ab7c733c109c6095f1277f619b9996f581fc3689633c33016f34739ddf" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.216805 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-skmxt" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.636279 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-njj2l"] Mar 12 15:03:17 crc kubenswrapper[4778]: E0312 15:03:17.636862 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434eb649-8d17-4d84-977f-a1907290d0f4" containerName="container-00" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.636875 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="434eb649-8d17-4d84-977f-a1907290d0f4" containerName="container-00" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.637066 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="434eb649-8d17-4d84-977f-a1907290d0f4" containerName="container-00" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.637708 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.675981 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-host\") pod \"crc-debug-njj2l\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.676046 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-576qw\" (UniqueName: \"kubernetes.io/projected/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-kube-api-access-576qw\") pod \"crc-debug-njj2l\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.778231 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-host\") pod \"crc-debug-njj2l\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.778295 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-576qw\" (UniqueName: \"kubernetes.io/projected/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-kube-api-access-576qw\") pod \"crc-debug-njj2l\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.778845 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-host\") pod \"crc-debug-njj2l\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.795702 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-576qw\" (UniqueName: \"kubernetes.io/projected/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-kube-api-access-576qw\") pod \"crc-debug-njj2l\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: I0312 15:03:17.958720 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:17 crc kubenswrapper[4778]: W0312 15:03:17.998316 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b5de73b_7d89_4665_bd2a_c0efc03f4e1d.slice/crio-4784c68754a85c758318ccd5a6b608473f5e1fd2b1db1f7dffff85a1bff83cbd WatchSource:0}: Error finding container 4784c68754a85c758318ccd5a6b608473f5e1fd2b1db1f7dffff85a1bff83cbd: Status 404 returned error can't find the container with id 4784c68754a85c758318ccd5a6b608473f5e1fd2b1db1f7dffff85a1bff83cbd Mar 12 15:03:18 crc kubenswrapper[4778]: I0312 15:03:18.225265 4778 generic.go:334] "Generic (PLEG): container finished" podID="4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" containerID="74340f2038c644c2a2c001699df4f77fd8e1cf73ce4885bad06a1749c4f74a6f" exitCode=0 Mar 12 15:03:18 crc kubenswrapper[4778]: I0312 15:03:18.225306 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-njj2l" event={"ID":"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d","Type":"ContainerDied","Data":"74340f2038c644c2a2c001699df4f77fd8e1cf73ce4885bad06a1749c4f74a6f"} Mar 12 15:03:18 crc kubenswrapper[4778]: I0312 15:03:18.225330 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/crc-debug-njj2l" event={"ID":"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d","Type":"ContainerStarted","Data":"4784c68754a85c758318ccd5a6b608473f5e1fd2b1db1f7dffff85a1bff83cbd"} Mar 12 15:03:18 crc kubenswrapper[4778]: I0312 15:03:18.269407 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434eb649-8d17-4d84-977f-a1907290d0f4" path="/var/lib/kubelet/pods/434eb649-8d17-4d84-977f-a1907290d0f4/volumes" Mar 12 15:03:18 crc kubenswrapper[4778]: I0312 15:03:18.270059 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-njj2l"] Mar 12 15:03:18 crc kubenswrapper[4778]: I0312 15:03:18.271649 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rkpvq/crc-debug-njj2l"] Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.334940 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.508996 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-576qw\" (UniqueName: \"kubernetes.io/projected/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-kube-api-access-576qw\") pod \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.509334 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-host\") pod \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\" (UID: \"4b5de73b-7d89-4665-bd2a-c0efc03f4e1d\") " Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.509563 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-host" (OuterVolumeSpecName: "host") pod "4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" (UID: "4b5de73b-7d89-4665-bd2a-c0efc03f4e1d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.509908 4778 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-host\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.514117 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-kube-api-access-576qw" (OuterVolumeSpecName: "kube-api-access-576qw") pod "4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" (UID: "4b5de73b-7d89-4665-bd2a-c0efc03f4e1d"). InnerVolumeSpecName "kube-api-access-576qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:03:19 crc kubenswrapper[4778]: I0312 15:03:19.611799 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-576qw\" (UniqueName: \"kubernetes.io/projected/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d-kube-api-access-576qw\") on node \"crc\" DevicePath \"\"" Mar 12 15:03:20 crc kubenswrapper[4778]: I0312 15:03:20.252957 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4784c68754a85c758318ccd5a6b608473f5e1fd2b1db1f7dffff85a1bff83cbd" Mar 12 15:03:20 crc kubenswrapper[4778]: I0312 15:03:20.253022 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/crc-debug-njj2l" Mar 12 15:03:20 crc kubenswrapper[4778]: I0312 15:03:20.271621 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" path="/var/lib/kubelet/pods/4b5de73b-7d89-4665-bd2a-c0efc03f4e1d/volumes" Mar 12 15:03:40 crc kubenswrapper[4778]: I0312 15:03:40.956086 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lvp8p"] Mar 12 15:03:40 crc kubenswrapper[4778]: E0312 15:03:40.956939 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" containerName="container-00" Mar 12 15:03:40 crc kubenswrapper[4778]: I0312 15:03:40.956952 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" containerName="container-00" Mar 12 15:03:40 crc kubenswrapper[4778]: I0312 15:03:40.957160 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b5de73b-7d89-4665-bd2a-c0efc03f4e1d" containerName="container-00" Mar 12 15:03:40 crc kubenswrapper[4778]: I0312 15:03:40.962386 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:40 crc kubenswrapper[4778]: I0312 15:03:40.973317 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvp8p"] Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.071788 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca67e14c-855d-473a-99b0-fe9dabb57916-catalog-content\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.071919 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca67e14c-855d-473a-99b0-fe9dabb57916-utilities\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.071962 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb7xc\" (UniqueName: \"kubernetes.io/projected/ca67e14c-855d-473a-99b0-fe9dabb57916-kube-api-access-sb7xc\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.173220 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca67e14c-855d-473a-99b0-fe9dabb57916-utilities\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.173525 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb7xc\" (UniqueName: \"kubernetes.io/projected/ca67e14c-855d-473a-99b0-fe9dabb57916-kube-api-access-sb7xc\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.173694 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca67e14c-855d-473a-99b0-fe9dabb57916-catalog-content\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.173836 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca67e14c-855d-473a-99b0-fe9dabb57916-utilities\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.174166 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca67e14c-855d-473a-99b0-fe9dabb57916-catalog-content\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.200123 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb7xc\" (UniqueName: \"kubernetes.io/projected/ca67e14c-855d-473a-99b0-fe9dabb57916-kube-api-access-sb7xc\") pod \"redhat-operators-lvp8p\" (UID: \"ca67e14c-855d-473a-99b0-fe9dabb57916\") " pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.341365 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:03:41 crc kubenswrapper[4778]: I0312 15:03:41.845298 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvp8p"] Mar 12 15:03:42 crc kubenswrapper[4778]: I0312 15:03:42.484406 4778 generic.go:334] "Generic (PLEG): container finished" podID="ca67e14c-855d-473a-99b0-fe9dabb57916" containerID="f5cf71f2c30496fc349cd115c9d22054161bef05f8a6c2dac0d3f20c006fccc5" exitCode=0 Mar 12 15:03:42 crc kubenswrapper[4778]: I0312 15:03:42.484487 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvp8p" event={"ID":"ca67e14c-855d-473a-99b0-fe9dabb57916","Type":"ContainerDied","Data":"f5cf71f2c30496fc349cd115c9d22054161bef05f8a6c2dac0d3f20c006fccc5"} Mar 12 15:03:42 crc kubenswrapper[4778]: I0312 15:03:42.484748 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvp8p" event={"ID":"ca67e14c-855d-473a-99b0-fe9dabb57916","Type":"ContainerStarted","Data":"5d37b81f2935c5aba929232694de9ae2f8c860dc2b2b539291da26409c61717c"} Mar 12 15:03:42 crc kubenswrapper[4778]: I0312 15:03:42.487799 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 15:03:48 crc kubenswrapper[4778]: I0312 15:03:48.801823 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-86cb765474-5pq5z_6bd172c5-383f-4273-98a5-2c92223dc765/barbican-api/0.log" Mar 12 15:03:48 crc kubenswrapper[4778]: I0312 15:03:48.953664 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-86cb765474-5pq5z_6bd172c5-383f-4273-98a5-2c92223dc765/barbican-api-log/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.065506 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65c9994dfd-xznqh_8ee1f546-8428-4b23-93e4-b8370fd4224b/barbican-keystone-listener/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.188637 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7dcf9787-ngc87_d505bb59-3c9e-4cfa-891c-c8e0068e2567/barbican-worker/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.336503 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65c9994dfd-xznqh_8ee1f546-8428-4b23-93e4-b8370fd4224b/barbican-keystone-listener-log/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.345167 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7dcf9787-ngc87_d505bb59-3c9e-4cfa-891c-c8e0068e2567/barbican-worker-log/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.476689 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx_b99627a8-43d8-4f7d-90f7-530eda3c2213/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.577414 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/ceilometer-central-agent/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.714064 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/proxy-httpd/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.762218 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/sg-core/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.805794 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/ceilometer-notification-agent/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.970952 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_99f72014-50e8-4dd4-9764-1b2c7d546b30/cinder-api/0.log" Mar 12 15:03:49 crc kubenswrapper[4778]: I0312 15:03:49.995240 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_39ee2404-53a8-4598-8c4b-c3a34fbf3480/cinder-scheduler/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.026674 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_99f72014-50e8-4dd4-9764-1b2c7d546b30/cinder-api-log/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.204548 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_39ee2404-53a8-4598-8c4b-c3a34fbf3480/probe/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.222774 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4szjl_5c5541f3-fb44-476b-91c2-b07dffe50894/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.437685 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6_36bb4acd-fab3-4998-a8cd-a6ebcc800fc8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.501103 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/init/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.668451 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/init/0.log" Mar 12 15:03:50 crc kubenswrapper[4778]: I0312 15:03:50.884589 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-2xksx_96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.091422 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_81c1a05c-5642-43d4-8a7b-229330168332/glance-httpd/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.121847 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_81c1a05c-5642-43d4-8a7b-229330168332/glance-log/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.207067 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/dnsmasq-dns/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.361341 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7fa757af-1c91-4b93-8916-5bbd99b8522e/glance-httpd/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.374050 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7fa757af-1c91-4b93-8916-5bbd99b8522e/glance-log/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.493923 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bngcx_f69e6cfe-f7c2-4127-b4df-710725c52227/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:51 crc kubenswrapper[4778]: I0312 15:03:51.571566 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g252n_29f8609b-4a3b-42ba-9450-a2b633bb4c2c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:52 crc kubenswrapper[4778]: I0312 15:03:52.098984 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555401-vjgkl_e4df6927-3452-4b36-b59a-a1fdcd4272a4/keystone-cron/0.log" Mar 12 15:03:52 crc kubenswrapper[4778]: I0312 15:03:52.341790 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555461-lmqk9_ebdf3274-70cb-4083-bf12-5d1038a9b7ba/keystone-cron/0.log" Mar 12 15:03:52 crc kubenswrapper[4778]: I0312 15:03:52.605654 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_51f24fcd-aff5-4785-abf7-4936180cee78/kube-state-metrics/0.log" Mar 12 15:03:52 crc kubenswrapper[4778]: I0312 15:03:52.931171 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8_8713b951-b516-42bd-9286-4343e5bcc955/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:53 crc kubenswrapper[4778]: I0312 15:03:53.247154 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-z4h9m_16dea17b-eaa4-4bbf-8895-c077b3e28d66/keystone-api/0.log" Mar 12 15:03:53 crc kubenswrapper[4778]: I0312 15:03:53.289166 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-6lrlq_a56bb599-f10d-4564-b6bf-48128dc2c7f1/keystone-api/0.log" Mar 12 15:03:54 crc kubenswrapper[4778]: I0312 15:03:54.028759 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-dggmh_7596a69e-33c9-4a2b-89fc-e4c41252b3fd/neutron-httpd/0.log" Mar 12 15:03:54 crc kubenswrapper[4778]: I0312 15:03:54.352310 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-zx97x_8a67d4b7-d8eb-40f4-b51d-62e92c6042c1/neutron-httpd/0.log" Mar 12 15:03:54 crc kubenswrapper[4778]: I0312 15:03:54.478509 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg_5cc410de-5b42-44d1-8b29-37161475730e/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:03:57 crc kubenswrapper[4778]: I0312 15:03:57.836449 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4/nova-api-log/0.log" Mar 12 15:03:59 crc kubenswrapper[4778]: I0312 15:03:59.318409 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-zx97x_8a67d4b7-d8eb-40f4-b51d-62e92c6042c1/neutron-api/0.log" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.168088 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555464-rt6fz"] Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.170421 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.172661 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.172855 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.174143 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.192623 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555464-rt6fz"] Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.269813 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pv9d\" (UniqueName: \"kubernetes.io/projected/bdc35b0a-1b16-4db8-adef-8a6afd6ae934-kube-api-access-6pv9d\") pod \"auto-csr-approver-29555464-rt6fz\" (UID: \"bdc35b0a-1b16-4db8-adef-8a6afd6ae934\") " pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.371772 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pv9d\" (UniqueName: \"kubernetes.io/projected/bdc35b0a-1b16-4db8-adef-8a6afd6ae934-kube-api-access-6pv9d\") pod \"auto-csr-approver-29555464-rt6fz\" (UID: \"bdc35b0a-1b16-4db8-adef-8a6afd6ae934\") " pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.404947 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pv9d\" (UniqueName: \"kubernetes.io/projected/bdc35b0a-1b16-4db8-adef-8a6afd6ae934-kube-api-access-6pv9d\") pod \"auto-csr-approver-29555464-rt6fz\" (UID: \"bdc35b0a-1b16-4db8-adef-8a6afd6ae934\") " pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.426157 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4/nova-api-api/0.log" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.487884 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:00 crc kubenswrapper[4778]: I0312 15:04:00.661130 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvp8p" event={"ID":"ca67e14c-855d-473a-99b0-fe9dabb57916","Type":"ContainerStarted","Data":"91ee2a929eebacad27622c766c9b3f9578ff1372b845836bc89b764f83c342a3"} Mar 12 15:04:01 crc kubenswrapper[4778]: I0312 15:04:01.267841 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555464-rt6fz"] Mar 12 15:04:01 crc kubenswrapper[4778]: I0312 15:04:01.429469 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-dggmh_7596a69e-33c9-4a2b-89fc-e4c41252b3fd/neutron-api/0.log" Mar 12 15:04:01 crc kubenswrapper[4778]: I0312 15:04:01.682710 4778 generic.go:334] "Generic (PLEG): container finished" podID="ca67e14c-855d-473a-99b0-fe9dabb57916" containerID="91ee2a929eebacad27622c766c9b3f9578ff1372b845836bc89b764f83c342a3" exitCode=0 Mar 12 15:04:01 crc kubenswrapper[4778]: I0312 15:04:01.682795 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvp8p" event={"ID":"ca67e14c-855d-473a-99b0-fe9dabb57916","Type":"ContainerDied","Data":"91ee2a929eebacad27622c766c9b3f9578ff1372b845836bc89b764f83c342a3"} Mar 12 15:04:01 crc kubenswrapper[4778]: I0312 15:04:01.684639 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" event={"ID":"bdc35b0a-1b16-4db8-adef-8a6afd6ae934","Type":"ContainerStarted","Data":"1d4f2fccef9895e998cbb789ad29a604968e2a751b694f840b5450a45509483f"} Mar 12 15:04:01 crc kubenswrapper[4778]: I0312 15:04:01.896557 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_929bb450-949d-4f4f-9c21-de6c3fe32927/nova-cell0-conductor-conductor/0.log" Mar 12 15:04:02 crc kubenswrapper[4778]: I0312 15:04:02.113503 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1466aea3-fa10-49a6-a254-a96a52091aca/nova-cell1-conductor-conductor/0.log" Mar 12 15:04:02 crc kubenswrapper[4778]: I0312 15:04:02.384111 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-1_f0341d80-4327-4c9e-bc11-0cddbc6eab66/nova-api-log/0.log" Mar 12 15:04:02 crc kubenswrapper[4778]: I0312 15:04:02.669783 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-metadata-0_c289a520-78eb-433f-b7a4-0c03be917c18/nova-cell1-metadata-log/0.log" Mar 12 15:04:02 crc kubenswrapper[4778]: I0312 15:04:02.983697 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-1_f0341d80-4327-4c9e-bc11-0cddbc6eab66/nova-api-api/0.log" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.139247 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7/nova-cell1-novncproxy-novncproxy/0.log" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.391031 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5tw6s_6ed77f87-e6b2-4c7a-8b0e-003106200dc8/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.656612 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-metadata-0_c289a520-78eb-433f-b7a4-0c03be917c18/nova-cell1-metadata-metadata/0.log" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.672582 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/mysql-bootstrap/0.log" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.701327 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvp8p" event={"ID":"ca67e14c-855d-473a-99b0-fe9dabb57916","Type":"ContainerStarted","Data":"3b4aa015c90f98a02842f429fad0b0b116679bfae043c37529385708f67f2ddd"} Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.702978 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" event={"ID":"bdc35b0a-1b16-4db8-adef-8a6afd6ae934","Type":"ContainerStarted","Data":"13a5daaa89f4db1da5ea953e47f7efe223f79130e65c2950cad2f908b81e7d1b"} Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.733628 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lvp8p" podStartSLOduration=2.893848084 podStartE2EDuration="23.733602018s" podCreationTimestamp="2026-03-12 15:03:40 +0000 UTC" firstStartedPulling="2026-03-12 15:03:42.487481667 +0000 UTC m=+6840.936177073" lastFinishedPulling="2026-03-12 15:04:03.327235611 +0000 UTC m=+6861.775931007" observedRunningTime="2026-03-12 15:04:03.717698836 +0000 UTC m=+6862.166394232" watchObservedRunningTime="2026-03-12 15:04:03.733602018 +0000 UTC m=+6862.182297414" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.735084 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" podStartSLOduration=1.637060443 podStartE2EDuration="3.73507896s" podCreationTimestamp="2026-03-12 15:04:00 +0000 UTC" firstStartedPulling="2026-03-12 15:04:01.27234767 +0000 UTC m=+6859.721043066" lastFinishedPulling="2026-03-12 15:04:03.370366187 +0000 UTC m=+6861.819061583" observedRunningTime="2026-03-12 15:04:03.733023992 +0000 UTC m=+6862.181719398" watchObservedRunningTime="2026-03-12 15:04:03.73507896 +0000 UTC m=+6862.183774346" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.900275 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/mysql-bootstrap/0.log" Mar 12 15:04:03 crc kubenswrapper[4778]: I0312 15:04:03.954873 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/galera/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.142069 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/mysql-bootstrap/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.377692 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/mysql-bootstrap/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.475148 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/galera/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.532390 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f613745b-fe33-4918-9e0a-da2a59c55e33/nova-scheduler-scheduler/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.630964 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_856cd6d1-db21-4503-94d7-cbf27ca96cc2/openstackclient/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.713815 4778 generic.go:334] "Generic (PLEG): container finished" podID="bdc35b0a-1b16-4db8-adef-8a6afd6ae934" containerID="13a5daaa89f4db1da5ea953e47f7efe223f79130e65c2950cad2f908b81e7d1b" exitCode=0 Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.714439 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" event={"ID":"bdc35b0a-1b16-4db8-adef-8a6afd6ae934","Type":"ContainerDied","Data":"13a5daaa89f4db1da5ea953e47f7efe223f79130e65c2950cad2f908b81e7d1b"} Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.922370 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vtt4z_a8484e5d-6f77-407c-81db-0d9b2a6b37fd/openstack-network-exporter/0.log" Mar 12 15:04:04 crc kubenswrapper[4778]: I0312 15:04:04.971828 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4wct6_3b8efd1e-884d-4963-b69f-04ede0a92267/ovn-controller/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.211828 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server-init/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.372875 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovs-vswitchd/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.428782 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server-init/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.431045 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.648138 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9lbdq_3c0a2200-506d-4ac3-b08c-9b3156c9e573/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.713926 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b25f9c9-784a-4a52-9bb3-02c6c4592702/openstack-network-exporter/0.log" Mar 12 15:04:05 crc kubenswrapper[4778]: I0312 15:04:05.752302 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b25f9c9-784a-4a52-9bb3-02c6c4592702/ovn-northd/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.054697 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7321e15e-673c-4e0d-80f8-6ac644c1940f/ovsdbserver-nb/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.098013 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7321e15e-673c-4e0d-80f8-6ac644c1940f/openstack-network-exporter/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.174708 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c951c6f-06fd-4793-a95b-26b5c1400d73/openstack-network-exporter/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.206786 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.296922 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c951c6f-06fd-4793-a95b-26b5c1400d73/ovsdbserver-sb/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.308622 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pv9d\" (UniqueName: \"kubernetes.io/projected/bdc35b0a-1b16-4db8-adef-8a6afd6ae934-kube-api-access-6pv9d\") pod \"bdc35b0a-1b16-4db8-adef-8a6afd6ae934\" (UID: \"bdc35b0a-1b16-4db8-adef-8a6afd6ae934\") " Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.321350 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc35b0a-1b16-4db8-adef-8a6afd6ae934-kube-api-access-6pv9d" (OuterVolumeSpecName: "kube-api-access-6pv9d") pod "bdc35b0a-1b16-4db8-adef-8a6afd6ae934" (UID: "bdc35b0a-1b16-4db8-adef-8a6afd6ae934"). InnerVolumeSpecName "kube-api-access-6pv9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.410523 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pv9d\" (UniqueName: \"kubernetes.io/projected/bdc35b0a-1b16-4db8-adef-8a6afd6ae934-kube-api-access-6pv9d\") on node \"crc\" DevicePath \"\"" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.683029 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/setup-container/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.732162 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" event={"ID":"bdc35b0a-1b16-4db8-adef-8a6afd6ae934","Type":"ContainerDied","Data":"1d4f2fccef9895e998cbb789ad29a604968e2a751b694f840b5450a45509483f"} Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.732205 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555464-rt6fz" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.732210 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d4f2fccef9895e998cbb789ad29a604968e2a751b694f840b5450a45509483f" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.812152 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d4d765698-l7bjx_267e7df2-d35c-45c4-af65-e8af31f8f6cf/placement-api/0.log" Mar 12 15:04:06 crc kubenswrapper[4778]: I0312 15:04:06.843687 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/setup-container/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.044734 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/rabbitmq/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.079619 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/setup-container/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.141384 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d4d765698-l7bjx_267e7df2-d35c-45c4-af65-e8af31f8f6cf/placement-log/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.277719 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555458-2kqth"] Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.287635 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555458-2kqth"] Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.320335 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/setup-container/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.447559 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/rabbitmq/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.452818 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc_43a3ffe4-8b64-4e26-b63a-5254a986e4a4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.690833 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc_bd7ac6b4-5600-45ce-b0ea-199dd4baefcb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:07 crc kubenswrapper[4778]: I0312 15:04:07.721113 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gt58t_b0bb06df-44bb-4939-9492-a6ad3d6b5368/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.010479 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8mmjm_c993b33e-6c36-4524-864a-65da461a8e0c/ssh-known-hosts-edpm-deployment/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.232399 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-77f887c49f-fw2qd_bbd76cb8-462f-4e60-b755-ef3170e70d11/proxy-server/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.265269 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa9dd73-2656-43b3-a6cb-634d312a166e" path="/var/lib/kubelet/pods/0fa9dd73-2656-43b3-a6cb-634d312a166e/volumes" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.282096 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5knbg_2edc2c90-f91e-402d-809c-514e9d8a5e04/swift-ring-rebalance/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.373327 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-77f887c49f-fw2qd_bbd76cb8-462f-4e60-b755-ef3170e70d11/proxy-httpd/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.552804 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-auditor/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.561545 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-reaper/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.690137 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-replicator/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.765881 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-server/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.802141 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-auditor/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.870173 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-replicator/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.965846 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-updater/0.log" Mar 12 15:04:08 crc kubenswrapper[4778]: I0312 15:04:08.966934 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-server/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.060595 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-auditor/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.151603 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-expirer/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.217435 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-replicator/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.221302 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-server/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.346905 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/rsync/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.351159 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-updater/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.459538 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/swift-recon-cron/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.692163 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s_2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.730873 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_74897d0a-ca7b-4589-bd4c-75910c2d491c/tempest-tests-tempest-tests-runner/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.797260 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_82246f69-2112-44e9-a783-a4a5926188b4/test-operator-logs-container/0.log" Mar 12 15:04:09 crc kubenswrapper[4778]: I0312 15:04:09.992713 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9glvr_41583476-38cd-4c0d-a05a-96ddc5b330ca/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:04:11 crc kubenswrapper[4778]: I0312 15:04:11.341799 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:04:11 crc kubenswrapper[4778]: I0312 15:04:11.342031 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:04:11 crc kubenswrapper[4778]: I0312 15:04:11.395582 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:04:11 crc kubenswrapper[4778]: I0312 15:04:11.827556 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lvp8p" Mar 12 15:04:11 crc kubenswrapper[4778]: I0312 15:04:11.981282 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvp8p"] Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.166746 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r99nz"] Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.166997 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r99nz" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="registry-server" containerID="cri-o://712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31" gracePeriod=2 Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.695227 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.789527 4778 generic.go:334] "Generic (PLEG): container finished" podID="89b39891-5207-4289-807f-57d00acb2937" containerID="712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31" exitCode=0 Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.791531 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99nz" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.792006 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99nz" event={"ID":"89b39891-5207-4289-807f-57d00acb2937","Type":"ContainerDied","Data":"712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31"} Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.792041 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99nz" event={"ID":"89b39891-5207-4289-807f-57d00acb2937","Type":"ContainerDied","Data":"d44d2bee1e2b4ddf99f45277d9dc014b3b21712ebf48e47cb48538e60ac5ff80"} Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.792067 4778 scope.go:117] "RemoveContainer" containerID="712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.820649 4778 scope.go:117] "RemoveContainer" containerID="805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.824731 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mccx\" (UniqueName: \"kubernetes.io/projected/89b39891-5207-4289-807f-57d00acb2937-kube-api-access-7mccx\") pod \"89b39891-5207-4289-807f-57d00acb2937\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.824933 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-utilities\") pod \"89b39891-5207-4289-807f-57d00acb2937\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.825012 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-catalog-content\") pod \"89b39891-5207-4289-807f-57d00acb2937\" (UID: \"89b39891-5207-4289-807f-57d00acb2937\") " Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.830734 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-utilities" (OuterVolumeSpecName: "utilities") pod "89b39891-5207-4289-807f-57d00acb2937" (UID: "89b39891-5207-4289-807f-57d00acb2937"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.835363 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b39891-5207-4289-807f-57d00acb2937-kube-api-access-7mccx" (OuterVolumeSpecName: "kube-api-access-7mccx") pod "89b39891-5207-4289-807f-57d00acb2937" (UID: "89b39891-5207-4289-807f-57d00acb2937"). InnerVolumeSpecName "kube-api-access-7mccx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.853275 4778 scope.go:117] "RemoveContainer" containerID="984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.927049 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.927079 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mccx\" (UniqueName: \"kubernetes.io/projected/89b39891-5207-4289-807f-57d00acb2937-kube-api-access-7mccx\") on node \"crc\" DevicePath \"\"" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.929411 4778 scope.go:117] "RemoveContainer" containerID="712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31" Mar 12 15:04:12 crc kubenswrapper[4778]: E0312 15:04:12.930541 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31\": container with ID starting with 712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31 not found: ID does not exist" containerID="712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.930572 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31"} err="failed to get container status \"712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31\": rpc error: code = NotFound desc = could not find container \"712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31\": container with ID starting with 712082342de67a11034de3ce859863eaaf1f71a829333a77a36e1df98eea2e31 not found: ID does not exist" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.930592 4778 scope.go:117] "RemoveContainer" containerID="805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e" Mar 12 15:04:12 crc kubenswrapper[4778]: E0312 15:04:12.930780 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e\": container with ID starting with 805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e not found: ID does not exist" containerID="805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.930799 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e"} err="failed to get container status \"805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e\": rpc error: code = NotFound desc = could not find container \"805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e\": container with ID starting with 805951d35b64df6e3a5d2f522d8ca4fce31a3962c15f8b2c7f8cc07a84f8dc1e not found: ID does not exist" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.930810 4778 scope.go:117] "RemoveContainer" containerID="984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b" Mar 12 15:04:12 crc kubenswrapper[4778]: E0312 15:04:12.931066 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b\": container with ID starting with 984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b not found: ID does not exist" containerID="984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.931086 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b"} err="failed to get container status \"984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b\": rpc error: code = NotFound desc = could not find container \"984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b\": container with ID starting with 984fb3456eea71c9cd7483dfcdb8376d81e856bd79aef84a885b305c1615885b not found: ID does not exist" Mar 12 15:04:12 crc kubenswrapper[4778]: I0312 15:04:12.989557 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89b39891-5207-4289-807f-57d00acb2937" (UID: "89b39891-5207-4289-807f-57d00acb2937"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:04:13 crc kubenswrapper[4778]: I0312 15:04:13.028508 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b39891-5207-4289-807f-57d00acb2937-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:04:13 crc kubenswrapper[4778]: I0312 15:04:13.129396 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r99nz"] Mar 12 15:04:13 crc kubenswrapper[4778]: I0312 15:04:13.144978 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r99nz"] Mar 12 15:04:14 crc kubenswrapper[4778]: I0312 15:04:14.263730 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b39891-5207-4289-807f-57d00acb2937" path="/var/lib/kubelet/pods/89b39891-5207-4289-807f-57d00acb2937/volumes" Mar 12 15:04:23 crc kubenswrapper[4778]: I0312 15:04:23.870762 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ec63cc68-6fde-419b-973c-91fc982e6a49/memcached/0.log" Mar 12 15:04:25 crc kubenswrapper[4778]: I0312 15:04:25.689286 4778 scope.go:117] "RemoveContainer" containerID="a5663c78d0886a072205a20f2510ea67c65b15026159b43c8bf3ff0037ce7434" Mar 12 15:04:28 crc kubenswrapper[4778]: I0312 15:04:28.558239 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:04:28 crc kubenswrapper[4778]: I0312 15:04:28.558660 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.556315 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/util/0.log" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.665541 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/util/0.log" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.720690 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/pull/0.log" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.761289 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/pull/0.log" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.932782 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/util/0.log" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.938676 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/pull/0.log" Mar 12 15:04:38 crc kubenswrapper[4778]: I0312 15:04:38.983893 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/extract/0.log" Mar 12 15:04:39 crc kubenswrapper[4778]: I0312 15:04:39.362306 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-9n6jv_ad531191-d7c5-4ef6-9929-3a5869751d98/manager/0.log" Mar 12 15:04:39 crc kubenswrapper[4778]: I0312 15:04:39.686642 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-gknp2_db7f6b97-2903-44bf-803f-c00c337400b9/manager/0.log" Mar 12 15:04:39 crc kubenswrapper[4778]: I0312 15:04:39.923913 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-b7tkm_e290c1ea-a39d-451e-a24b-17a2b61ff6f0/manager/0.log" Mar 12 15:04:40 crc kubenswrapper[4778]: I0312 15:04:40.135739 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-4jgt8_4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71/manager/0.log" Mar 12 15:04:40 crc kubenswrapper[4778]: I0312 15:04:40.773975 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-5d6qz_02bc06ca-f4e6-4fde-bd5d-882714d9652c/manager/0.log" Mar 12 15:04:40 crc kubenswrapper[4778]: I0312 15:04:40.853097 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-qb8s8_98a4cfbd-3037-48b5-9047-5d574dcc0aca/manager/0.log" Mar 12 15:04:41 crc kubenswrapper[4778]: I0312 15:04:41.223651 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7dxdh_7e02c37f-b9af-46c9-a743-03ead9b060db/manager/0.log" Mar 12 15:04:41 crc kubenswrapper[4778]: I0312 15:04:41.712023 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-pn8tk_5e38a4fd-95f8-437b-923b-eca33b1387e6/manager/0.log" Mar 12 15:04:41 crc kubenswrapper[4778]: I0312 15:04:41.733266 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-xm4cc_c8818ac0-af8b-42c9-a923-425fe79ed203/manager/0.log" Mar 12 15:04:41 crc kubenswrapper[4778]: I0312 15:04:41.931861 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-jlbft_2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f/manager/0.log" Mar 12 15:04:42 crc kubenswrapper[4778]: I0312 15:04:42.302303 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-dd2ft_076835c9-352b-4e40-80c4-3bce3bb80594/manager/0.log" Mar 12 15:04:42 crc kubenswrapper[4778]: I0312 15:04:42.504215 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-686d5f9fbd-vv9rc_d7288cc6-4247-4d03-bd37-9862243bf613/manager/0.log" Mar 12 15:04:42 crc kubenswrapper[4778]: I0312 15:04:42.582055 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-cdgg9_1a01d06c-be6f-45de-a22d-c8f1058a3a84/manager/0.log" Mar 12 15:04:42 crc kubenswrapper[4778]: I0312 15:04:42.782502 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6_4f7d316e-6896-4f84-8423-6f79778c1c6b/manager/0.log" Mar 12 15:04:43 crc kubenswrapper[4778]: I0312 15:04:43.106202 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-5bc4df7446-x9bsl_34bbdc16-4518-4ee5-9a70-3cedcc5f0159/operator/0.log" Mar 12 15:04:43 crc kubenswrapper[4778]: I0312 15:04:43.253263 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-b2fsv_748546a6-1355-470f-b8d0-de395cf3f681/registry-server/0.log" Mar 12 15:04:43 crc kubenswrapper[4778]: I0312 15:04:43.418301 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-bbgmb_8d38fd7e-6fa1-4b0c-9c82-9c57290c7837/manager/0.log" Mar 12 15:04:43 crc kubenswrapper[4778]: I0312 15:04:43.544491 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-wvpf8_52524252-25bd-49e5-822e-3d4668aff2f9/manager/0.log" Mar 12 15:04:43 crc kubenswrapper[4778]: I0312 15:04:43.756444 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-shf7b_034f39d8-a33e-4e37-bcde-51fb22debdd1/operator/0.log" Mar 12 15:04:43 crc kubenswrapper[4778]: I0312 15:04:43.860798 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-84mps_64a36384-f2e6-4077-b2ca-de2a6ce6ea06/manager/0.log" Mar 12 15:04:44 crc kubenswrapper[4778]: I0312 15:04:44.108810 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-gfv5z_6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c/manager/0.log" Mar 12 15:04:44 crc kubenswrapper[4778]: I0312 15:04:44.388434 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-pcfrz_ed9b9271-4ae9-440a-9411-15d46267106e/manager/0.log" Mar 12 15:04:44 crc kubenswrapper[4778]: I0312 15:04:44.549836 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-2tjsk_8c02ecb8-0e15-4672-823a-c4437ca5bf8c/manager/0.log" Mar 12 15:04:44 crc kubenswrapper[4778]: I0312 15:04:44.729511 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5785b7957-7vdgw_d0784623-5f08-4109-9c7e-0a329210ce07/manager/0.log" Mar 12 15:04:49 crc kubenswrapper[4778]: I0312 15:04:49.515662 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-6h2c2_ffb8a1f4-4533-4368-a900-95d37fe1d3ad/manager/0.log" Mar 12 15:04:58 crc kubenswrapper[4778]: I0312 15:04:58.558286 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:04:58 crc kubenswrapper[4778]: I0312 15:04:58.559046 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:05:06 crc kubenswrapper[4778]: I0312 15:05:06.925945 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zkrqr_f799c7e9-1c31-40bc-9ece-06a086683a98/control-plane-machine-set-operator/0.log" Mar 12 15:05:07 crc kubenswrapper[4778]: I0312 15:05:07.114814 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-242cb_e2967620-e2ce-4763-8a6c-e5a37f3a1f98/machine-api-operator/0.log" Mar 12 15:05:07 crc kubenswrapper[4778]: I0312 15:05:07.160383 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-242cb_e2967620-e2ce-4763-8a6c-e5a37f3a1f98/kube-rbac-proxy/0.log" Mar 12 15:05:20 crc kubenswrapper[4778]: I0312 15:05:20.973009 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-2774s_92b29110-f478-42b5-9a5f-c9330a3973b2/cert-manager-controller/0.log" Mar 12 15:05:21 crc kubenswrapper[4778]: I0312 15:05:21.258526 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-jxs4g_804d0b09-6fab-4277-936a-5e0324d76b3e/cert-manager-cainjector/0.log" Mar 12 15:05:21 crc kubenswrapper[4778]: I0312 15:05:21.272000 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-ffh2x_45da07c5-bccb-4433-aa38-d9d2894f1b09/cert-manager-webhook/0.log" Mar 12 15:05:28 crc kubenswrapper[4778]: I0312 15:05:28.557605 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:05:28 crc kubenswrapper[4778]: I0312 15:05:28.558089 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:05:28 crc kubenswrapper[4778]: I0312 15:05:28.558144 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 15:05:28 crc kubenswrapper[4778]: I0312 15:05:28.558888 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"009e612c3693545ba4a1988aa00993d05612427ec6eb485b08b455b35968f1ab"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 15:05:28 crc kubenswrapper[4778]: I0312 15:05:28.558956 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://009e612c3693545ba4a1988aa00993d05612427ec6eb485b08b455b35968f1ab" gracePeriod=600 Mar 12 15:05:29 crc kubenswrapper[4778]: I0312 15:05:29.508568 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="009e612c3693545ba4a1988aa00993d05612427ec6eb485b08b455b35968f1ab" exitCode=0 Mar 12 15:05:29 crc kubenswrapper[4778]: I0312 15:05:29.508648 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"009e612c3693545ba4a1988aa00993d05612427ec6eb485b08b455b35968f1ab"} Mar 12 15:05:29 crc kubenswrapper[4778]: I0312 15:05:29.509119 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930"} Mar 12 15:05:29 crc kubenswrapper[4778]: I0312 15:05:29.509146 4778 scope.go:117] "RemoveContainer" containerID="505b7ca3387092da837254cfad64e23448af9dbba84199bbb89de928d39d31e3" Mar 12 15:05:36 crc kubenswrapper[4778]: I0312 15:05:36.267462 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-jbxx4_af2d568b-9719-4da9-b0e8-e28d314ed860/nmstate-console-plugin/0.log" Mar 12 15:05:36 crc kubenswrapper[4778]: I0312 15:05:36.437926 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rbsjl_d8309ffe-a26c-44a8-84e2-7b7ec10982a8/nmstate-handler/0.log" Mar 12 15:05:36 crc kubenswrapper[4778]: I0312 15:05:36.504605 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-b2s5h_7855d7b1-c7cf-4b63-9313-051a391fcf43/nmstate-metrics/0.log" Mar 12 15:05:36 crc kubenswrapper[4778]: I0312 15:05:36.509127 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-b2s5h_7855d7b1-c7cf-4b63-9313-051a391fcf43/kube-rbac-proxy/0.log" Mar 12 15:05:36 crc kubenswrapper[4778]: I0312 15:05:36.638436 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-hxzd6_fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3/nmstate-operator/0.log" Mar 12 15:05:36 crc kubenswrapper[4778]: I0312 15:05:36.920957 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-94rbc_ef796a94-b10d-4d18-ae88-f64bc3a6b87d/nmstate-webhook/0.log" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.150136 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555466-7wlhj"] Mar 12 15:06:00 crc kubenswrapper[4778]: E0312 15:06:00.151332 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc35b0a-1b16-4db8-adef-8a6afd6ae934" containerName="oc" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.151352 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc35b0a-1b16-4db8-adef-8a6afd6ae934" containerName="oc" Mar 12 15:06:00 crc kubenswrapper[4778]: E0312 15:06:00.151381 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="registry-server" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.151390 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="registry-server" Mar 12 15:06:00 crc kubenswrapper[4778]: E0312 15:06:00.151410 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="extract-utilities" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.151419 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="extract-utilities" Mar 12 15:06:00 crc kubenswrapper[4778]: E0312 15:06:00.151450 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="extract-content" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.151458 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="extract-content" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.151689 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b39891-5207-4289-807f-57d00acb2937" containerName="registry-server" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.151715 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc35b0a-1b16-4db8-adef-8a6afd6ae934" containerName="oc" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.152697 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.157855 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.158307 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.158519 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.163891 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555466-7wlhj"] Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.266312 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxxck\" (UniqueName: \"kubernetes.io/projected/c58e0f99-4ece-49ec-9c47-b82055df7d48-kube-api-access-bxxck\") pod \"auto-csr-approver-29555466-7wlhj\" (UID: \"c58e0f99-4ece-49ec-9c47-b82055df7d48\") " pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.368757 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxxck\" (UniqueName: \"kubernetes.io/projected/c58e0f99-4ece-49ec-9c47-b82055df7d48-kube-api-access-bxxck\") pod \"auto-csr-approver-29555466-7wlhj\" (UID: \"c58e0f99-4ece-49ec-9c47-b82055df7d48\") " pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.387335 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxxck\" (UniqueName: \"kubernetes.io/projected/c58e0f99-4ece-49ec-9c47-b82055df7d48-kube-api-access-bxxck\") pod \"auto-csr-approver-29555466-7wlhj\" (UID: \"c58e0f99-4ece-49ec-9c47-b82055df7d48\") " pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.475695 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:00 crc kubenswrapper[4778]: I0312 15:06:00.920829 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555466-7wlhj"] Mar 12 15:06:01 crc kubenswrapper[4778]: I0312 15:06:01.836836 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" event={"ID":"c58e0f99-4ece-49ec-9c47-b82055df7d48","Type":"ContainerStarted","Data":"65f40878aba2160868688af649b2ce7d73b76d4df721a61bcd9c29b8d55924bf"} Mar 12 15:06:02 crc kubenswrapper[4778]: I0312 15:06:02.852460 4778 generic.go:334] "Generic (PLEG): container finished" podID="c58e0f99-4ece-49ec-9c47-b82055df7d48" containerID="8984f879d02ecea61666d68e6857174dd681c26238a5c27d5a617cc7dccda3db" exitCode=0 Mar 12 15:06:02 crc kubenswrapper[4778]: I0312 15:06:02.852949 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" event={"ID":"c58e0f99-4ece-49ec-9c47-b82055df7d48","Type":"ContainerDied","Data":"8984f879d02ecea61666d68e6857174dd681c26238a5c27d5a617cc7dccda3db"} Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.266899 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.355467 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxxck\" (UniqueName: \"kubernetes.io/projected/c58e0f99-4ece-49ec-9c47-b82055df7d48-kube-api-access-bxxck\") pod \"c58e0f99-4ece-49ec-9c47-b82055df7d48\" (UID: \"c58e0f99-4ece-49ec-9c47-b82055df7d48\") " Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.361011 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58e0f99-4ece-49ec-9c47-b82055df7d48-kube-api-access-bxxck" (OuterVolumeSpecName: "kube-api-access-bxxck") pod "c58e0f99-4ece-49ec-9c47-b82055df7d48" (UID: "c58e0f99-4ece-49ec-9c47-b82055df7d48"). InnerVolumeSpecName "kube-api-access-bxxck". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.458215 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxxck\" (UniqueName: \"kubernetes.io/projected/c58e0f99-4ece-49ec-9c47-b82055df7d48-kube-api-access-bxxck\") on node \"crc\" DevicePath \"\"" Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.873799 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" event={"ID":"c58e0f99-4ece-49ec-9c47-b82055df7d48","Type":"ContainerDied","Data":"65f40878aba2160868688af649b2ce7d73b76d4df721a61bcd9c29b8d55924bf"} Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.874076 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65f40878aba2160868688af649b2ce7d73b76d4df721a61bcd9c29b8d55924bf" Mar 12 15:06:04 crc kubenswrapper[4778]: I0312 15:06:04.873835 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555466-7wlhj" Mar 12 15:06:05 crc kubenswrapper[4778]: I0312 15:06:05.346943 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555460-6bwr2"] Mar 12 15:06:05 crc kubenswrapper[4778]: I0312 15:06:05.356747 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555460-6bwr2"] Mar 12 15:06:06 crc kubenswrapper[4778]: I0312 15:06:06.264176 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7baca351-722e-4d7e-972e-04513fae6e0b" path="/var/lib/kubelet/pods/7baca351-722e-4d7e-972e-04513fae6e0b/volumes" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.313974 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mnjql_14351deb-3286-4464-8eac-6bb116a9ebce/kube-rbac-proxy/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.402503 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mnjql_14351deb-3286-4464-8eac-6bb116a9ebce/controller/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.489330 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-x2n7f_2f214887-d638-42fa-aa86-1518cfae600d/frr-k8s-webhook-server/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.569820 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.761910 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.803886 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.804567 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.808262 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:06:07 crc kubenswrapper[4778]: I0312 15:06:07.983580 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.031079 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.031445 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.081079 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.292944 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.296527 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/controller/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.325079 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.327211 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.490302 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/kube-rbac-proxy-frr/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.505001 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/kube-rbac-proxy/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.534786 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/frr-metrics/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.736759 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/reloader/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.772304 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-54d5c4b6c7-gh4lx_a5a6d344-0a75-422d-acd9-fe8887b03110/manager/0.log" Mar 12 15:06:08 crc kubenswrapper[4778]: I0312 15:06:08.924178 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68f5db54d6-zstmq_6ac207b6-1710-47af-8fe9-b0c3adbce0ab/webhook-server/0.log" Mar 12 15:06:09 crc kubenswrapper[4778]: I0312 15:06:09.111606 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k7nvk_f2e1d11e-8f27-498d-8d45-ac0e14a796fe/kube-rbac-proxy/0.log" Mar 12 15:06:09 crc kubenswrapper[4778]: I0312 15:06:09.680663 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k7nvk_f2e1d11e-8f27-498d-8d45-ac0e14a796fe/speaker/0.log" Mar 12 15:06:10 crc kubenswrapper[4778]: I0312 15:06:10.709840 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/frr/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.260665 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/util/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.469811 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/util/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.486990 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/pull/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.553979 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/pull/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.770481 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/pull/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.775100 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/util/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.802922 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/extract/0.log" Mar 12 15:06:23 crc kubenswrapper[4778]: I0312 15:06:23.957649 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/util/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.140388 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/pull/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.156467 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/util/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.175375 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/pull/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.397088 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/pull/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.404213 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/extract/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.408632 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/util/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.613624 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-utilities/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.777360 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-utilities/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.786146 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-content/0.log" Mar 12 15:06:24 crc kubenswrapper[4778]: I0312 15:06:24.788319 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-content/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.007179 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-utilities/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.021887 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-content/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.211183 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-utilities/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.447473 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-utilities/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.454714 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-content/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.526116 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-content/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.734640 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-content/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.765702 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-utilities/0.log" Mar 12 15:06:25 crc kubenswrapper[4778]: I0312 15:06:25.807917 4778 scope.go:117] "RemoveContainer" containerID="64150eeb0f1f171e7d11ada7712192a8c533967a0e598d41c325a6422f027d7a" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.006474 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hvmk8_3b062c23-5acd-430d-aa6c-24b48a725594/marketplace-operator/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.223376 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/registry-server/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.269154 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-utilities/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.342025 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/registry-server/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.439733 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-content/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.479463 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-utilities/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.489383 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-content/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.652984 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-utilities/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.680415 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-content/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.890004 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/registry-server/0.log" Mar 12 15:06:26 crc kubenswrapper[4778]: I0312 15:06:26.910405 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-utilities/0.log" Mar 12 15:06:27 crc kubenswrapper[4778]: I0312 15:06:27.145789 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-utilities/0.log" Mar 12 15:06:27 crc kubenswrapper[4778]: I0312 15:06:27.145807 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-content/0.log" Mar 12 15:06:27 crc kubenswrapper[4778]: I0312 15:06:27.146855 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-content/0.log" Mar 12 15:06:27 crc kubenswrapper[4778]: I0312 15:06:27.336107 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-content/0.log" Mar 12 15:06:27 crc kubenswrapper[4778]: I0312 15:06:27.380165 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-utilities/0.log" Mar 12 15:06:27 crc kubenswrapper[4778]: I0312 15:06:27.453480 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/registry-server/0.log" Mar 12 15:06:45 crc kubenswrapper[4778]: E0312 15:06:45.498228 4778 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.32:54840->38.129.56.32:35979: write tcp 38.129.56.32:54840->38.129.56.32:35979: write: broken pipe Mar 12 15:07:28 crc kubenswrapper[4778]: I0312 15:07:28.558393 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:07:28 crc kubenswrapper[4778]: I0312 15:07:28.559100 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:07:58 crc kubenswrapper[4778]: I0312 15:07:58.557799 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:07:58 crc kubenswrapper[4778]: I0312 15:07:58.558376 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.166920 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555468-tpk68"] Mar 12 15:08:00 crc kubenswrapper[4778]: E0312 15:08:00.167930 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58e0f99-4ece-49ec-9c47-b82055df7d48" containerName="oc" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.167962 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58e0f99-4ece-49ec-9c47-b82055df7d48" containerName="oc" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.168392 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58e0f99-4ece-49ec-9c47-b82055df7d48" containerName="oc" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.169539 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.172244 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.172674 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.174063 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.186856 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555468-tpk68"] Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.267213 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj55d\" (UniqueName: \"kubernetes.io/projected/1760b72d-ab0b-489f-b263-7279ce51dc5f-kube-api-access-zj55d\") pod \"auto-csr-approver-29555468-tpk68\" (UID: \"1760b72d-ab0b-489f-b263-7279ce51dc5f\") " pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.368682 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj55d\" (UniqueName: \"kubernetes.io/projected/1760b72d-ab0b-489f-b263-7279ce51dc5f-kube-api-access-zj55d\") pod \"auto-csr-approver-29555468-tpk68\" (UID: \"1760b72d-ab0b-489f-b263-7279ce51dc5f\") " pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.394780 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj55d\" (UniqueName: \"kubernetes.io/projected/1760b72d-ab0b-489f-b263-7279ce51dc5f-kube-api-access-zj55d\") pod \"auto-csr-approver-29555468-tpk68\" (UID: \"1760b72d-ab0b-489f-b263-7279ce51dc5f\") " pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:00 crc kubenswrapper[4778]: I0312 15:08:00.492389 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:01 crc kubenswrapper[4778]: I0312 15:08:01.011353 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555468-tpk68"] Mar 12 15:08:02 crc kubenswrapper[4778]: I0312 15:08:02.041233 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555468-tpk68" event={"ID":"1760b72d-ab0b-489f-b263-7279ce51dc5f","Type":"ContainerStarted","Data":"c2f2c23d12c7f8a6128ee9c89c036a933a8cc16f980be4d5025a97fae92ade2a"} Mar 12 15:08:03 crc kubenswrapper[4778]: I0312 15:08:03.053710 4778 generic.go:334] "Generic (PLEG): container finished" podID="1760b72d-ab0b-489f-b263-7279ce51dc5f" containerID="98229a921540253e11a1f90e715794eabbd7a1c19952afa6969c95cd62f6a069" exitCode=0 Mar 12 15:08:03 crc kubenswrapper[4778]: I0312 15:08:03.054152 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555468-tpk68" event={"ID":"1760b72d-ab0b-489f-b263-7279ce51dc5f","Type":"ContainerDied","Data":"98229a921540253e11a1f90e715794eabbd7a1c19952afa6969c95cd62f6a069"} Mar 12 15:08:04 crc kubenswrapper[4778]: I0312 15:08:04.384518 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:04 crc kubenswrapper[4778]: I0312 15:08:04.556379 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj55d\" (UniqueName: \"kubernetes.io/projected/1760b72d-ab0b-489f-b263-7279ce51dc5f-kube-api-access-zj55d\") pod \"1760b72d-ab0b-489f-b263-7279ce51dc5f\" (UID: \"1760b72d-ab0b-489f-b263-7279ce51dc5f\") " Mar 12 15:08:04 crc kubenswrapper[4778]: I0312 15:08:04.564540 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1760b72d-ab0b-489f-b263-7279ce51dc5f-kube-api-access-zj55d" (OuterVolumeSpecName: "kube-api-access-zj55d") pod "1760b72d-ab0b-489f-b263-7279ce51dc5f" (UID: "1760b72d-ab0b-489f-b263-7279ce51dc5f"). InnerVolumeSpecName "kube-api-access-zj55d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:08:04 crc kubenswrapper[4778]: I0312 15:08:04.658372 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj55d\" (UniqueName: \"kubernetes.io/projected/1760b72d-ab0b-489f-b263-7279ce51dc5f-kube-api-access-zj55d\") on node \"crc\" DevicePath \"\"" Mar 12 15:08:05 crc kubenswrapper[4778]: I0312 15:08:05.082960 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555468-tpk68" event={"ID":"1760b72d-ab0b-489f-b263-7279ce51dc5f","Type":"ContainerDied","Data":"c2f2c23d12c7f8a6128ee9c89c036a933a8cc16f980be4d5025a97fae92ade2a"} Mar 12 15:08:05 crc kubenswrapper[4778]: I0312 15:08:05.083393 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f2c23d12c7f8a6128ee9c89c036a933a8cc16f980be4d5025a97fae92ade2a" Mar 12 15:08:05 crc kubenswrapper[4778]: I0312 15:08:05.083052 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555468-tpk68" Mar 12 15:08:05 crc kubenswrapper[4778]: I0312 15:08:05.455275 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555462-dvtsm"] Mar 12 15:08:05 crc kubenswrapper[4778]: I0312 15:08:05.464562 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555462-dvtsm"] Mar 12 15:08:06 crc kubenswrapper[4778]: I0312 15:08:06.268330 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb23378-6e3e-4c63-919d-47ce1d17dd7b" path="/var/lib/kubelet/pods/cbb23378-6e3e-4c63-919d-47ce1d17dd7b/volumes" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.738801 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qcswd"] Mar 12 15:08:07 crc kubenswrapper[4778]: E0312 15:08:07.739628 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1760b72d-ab0b-489f-b263-7279ce51dc5f" containerName="oc" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.739652 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="1760b72d-ab0b-489f-b263-7279ce51dc5f" containerName="oc" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.739975 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="1760b72d-ab0b-489f-b263-7279ce51dc5f" containerName="oc" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.742500 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.773472 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qcswd"] Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.838924 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6cfr\" (UniqueName: \"kubernetes.io/projected/193a1938-83bf-48dc-abd9-ecc2f202db8d-kube-api-access-g6cfr\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.839012 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-catalog-content\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.839143 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-utilities\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.941042 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-utilities\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.941151 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6cfr\" (UniqueName: \"kubernetes.io/projected/193a1938-83bf-48dc-abd9-ecc2f202db8d-kube-api-access-g6cfr\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.941231 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-catalog-content\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.941632 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-utilities\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.941722 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-catalog-content\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:07 crc kubenswrapper[4778]: I0312 15:08:07.960489 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6cfr\" (UniqueName: \"kubernetes.io/projected/193a1938-83bf-48dc-abd9-ecc2f202db8d-kube-api-access-g6cfr\") pod \"certified-operators-qcswd\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:08 crc kubenswrapper[4778]: I0312 15:08:08.083428 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:08 crc kubenswrapper[4778]: I0312 15:08:08.371530 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qcswd"] Mar 12 15:08:08 crc kubenswrapper[4778]: W0312 15:08:08.374542 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod193a1938_83bf_48dc_abd9_ecc2f202db8d.slice/crio-f6806f156fc16e1a064496e9b8083fb2d3b7f16ce78f083b0ab2e5de6a315557 WatchSource:0}: Error finding container f6806f156fc16e1a064496e9b8083fb2d3b7f16ce78f083b0ab2e5de6a315557: Status 404 returned error can't find the container with id f6806f156fc16e1a064496e9b8083fb2d3b7f16ce78f083b0ab2e5de6a315557 Mar 12 15:08:09 crc kubenswrapper[4778]: I0312 15:08:09.130393 4778 generic.go:334] "Generic (PLEG): container finished" podID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerID="33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2" exitCode=0 Mar 12 15:08:09 crc kubenswrapper[4778]: I0312 15:08:09.130712 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerDied","Data":"33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2"} Mar 12 15:08:09 crc kubenswrapper[4778]: I0312 15:08:09.130817 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerStarted","Data":"f6806f156fc16e1a064496e9b8083fb2d3b7f16ce78f083b0ab2e5de6a315557"} Mar 12 15:08:11 crc kubenswrapper[4778]: I0312 15:08:11.158036 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerStarted","Data":"c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a"} Mar 12 15:08:12 crc kubenswrapper[4778]: I0312 15:08:12.169671 4778 generic.go:334] "Generic (PLEG): container finished" podID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerID="c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a" exitCode=0 Mar 12 15:08:12 crc kubenswrapper[4778]: I0312 15:08:12.169716 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerDied","Data":"c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a"} Mar 12 15:08:13 crc kubenswrapper[4778]: I0312 15:08:13.180408 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerStarted","Data":"5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6"} Mar 12 15:08:13 crc kubenswrapper[4778]: I0312 15:08:13.211069 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qcswd" podStartSLOduration=2.666458982 podStartE2EDuration="6.211049135s" podCreationTimestamp="2026-03-12 15:08:07 +0000 UTC" firstStartedPulling="2026-03-12 15:08:09.132985257 +0000 UTC m=+7107.581680653" lastFinishedPulling="2026-03-12 15:08:12.67757538 +0000 UTC m=+7111.126270806" observedRunningTime="2026-03-12 15:08:13.206875716 +0000 UTC m=+7111.655571112" watchObservedRunningTime="2026-03-12 15:08:13.211049135 +0000 UTC m=+7111.659744541" Mar 12 15:08:18 crc kubenswrapper[4778]: I0312 15:08:18.084530 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:18 crc kubenswrapper[4778]: I0312 15:08:18.086436 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:18 crc kubenswrapper[4778]: I0312 15:08:18.148813 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:18 crc kubenswrapper[4778]: I0312 15:08:18.282960 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:18 crc kubenswrapper[4778]: I0312 15:08:18.396459 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qcswd"] Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.255102 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qcswd" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="registry-server" containerID="cri-o://5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6" gracePeriod=2 Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.770272 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.917938 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6cfr\" (UniqueName: \"kubernetes.io/projected/193a1938-83bf-48dc-abd9-ecc2f202db8d-kube-api-access-g6cfr\") pod \"193a1938-83bf-48dc-abd9-ecc2f202db8d\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.918280 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-utilities\") pod \"193a1938-83bf-48dc-abd9-ecc2f202db8d\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.918439 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-catalog-content\") pod \"193a1938-83bf-48dc-abd9-ecc2f202db8d\" (UID: \"193a1938-83bf-48dc-abd9-ecc2f202db8d\") " Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.922402 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-utilities" (OuterVolumeSpecName: "utilities") pod "193a1938-83bf-48dc-abd9-ecc2f202db8d" (UID: "193a1938-83bf-48dc-abd9-ecc2f202db8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:08:20 crc kubenswrapper[4778]: I0312 15:08:20.926710 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193a1938-83bf-48dc-abd9-ecc2f202db8d-kube-api-access-g6cfr" (OuterVolumeSpecName: "kube-api-access-g6cfr") pod "193a1938-83bf-48dc-abd9-ecc2f202db8d" (UID: "193a1938-83bf-48dc-abd9-ecc2f202db8d"). InnerVolumeSpecName "kube-api-access-g6cfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.021954 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.022318 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6cfr\" (UniqueName: \"kubernetes.io/projected/193a1938-83bf-48dc-abd9-ecc2f202db8d-kube-api-access-g6cfr\") on node \"crc\" DevicePath \"\"" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.267206 4778 generic.go:334] "Generic (PLEG): container finished" podID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerID="5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6" exitCode=0 Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.267269 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcswd" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.267292 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerDied","Data":"5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6"} Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.268597 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcswd" event={"ID":"193a1938-83bf-48dc-abd9-ecc2f202db8d","Type":"ContainerDied","Data":"f6806f156fc16e1a064496e9b8083fb2d3b7f16ce78f083b0ab2e5de6a315557"} Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.268619 4778 scope.go:117] "RemoveContainer" containerID="5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.320092 4778 scope.go:117] "RemoveContainer" containerID="c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.349001 4778 scope.go:117] "RemoveContainer" containerID="33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.413876 4778 scope.go:117] "RemoveContainer" containerID="5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6" Mar 12 15:08:21 crc kubenswrapper[4778]: E0312 15:08:21.414212 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6\": container with ID starting with 5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6 not found: ID does not exist" containerID="5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.414245 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6"} err="failed to get container status \"5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6\": rpc error: code = NotFound desc = could not find container \"5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6\": container with ID starting with 5f7899fb9193c2e9377a242c9cfedeeacd5e89ca82449d566c5a464428b75dc6 not found: ID does not exist" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.414269 4778 scope.go:117] "RemoveContainer" containerID="c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a" Mar 12 15:08:21 crc kubenswrapper[4778]: E0312 15:08:21.414679 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a\": container with ID starting with c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a not found: ID does not exist" containerID="c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.414699 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a"} err="failed to get container status \"c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a\": rpc error: code = NotFound desc = could not find container \"c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a\": container with ID starting with c5a282a10d6fde57de226d764c03613a263de2ccb99a3386ebf22eb1e1993b4a not found: ID does not exist" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.414712 4778 scope.go:117] "RemoveContainer" containerID="33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2" Mar 12 15:08:21 crc kubenswrapper[4778]: E0312 15:08:21.415167 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2\": container with ID starting with 33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2 not found: ID does not exist" containerID="33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2" Mar 12 15:08:21 crc kubenswrapper[4778]: I0312 15:08:21.415209 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2"} err="failed to get container status \"33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2\": rpc error: code = NotFound desc = could not find container \"33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2\": container with ID starting with 33ad6411c9f8c1d5d5bb41623664b1079e456cd6577aad63ed94d81ea4af1da2 not found: ID does not exist" Mar 12 15:08:22 crc kubenswrapper[4778]: I0312 15:08:22.494591 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "193a1938-83bf-48dc-abd9-ecc2f202db8d" (UID: "193a1938-83bf-48dc-abd9-ecc2f202db8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:08:22 crc kubenswrapper[4778]: I0312 15:08:22.557683 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a1938-83bf-48dc-abd9-ecc2f202db8d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:08:22 crc kubenswrapper[4778]: I0312 15:08:22.808524 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qcswd"] Mar 12 15:08:22 crc kubenswrapper[4778]: I0312 15:08:22.818282 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qcswd"] Mar 12 15:08:24 crc kubenswrapper[4778]: I0312 15:08:24.275877 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" path="/var/lib/kubelet/pods/193a1938-83bf-48dc-abd9-ecc2f202db8d/volumes" Mar 12 15:08:25 crc kubenswrapper[4778]: I0312 15:08:25.939899 4778 scope.go:117] "RemoveContainer" containerID="bc1c69d732ac8380ce4ad84b76897a91373ec3edde2343f57d27f4105f4594eb" Mar 12 15:08:26 crc kubenswrapper[4778]: I0312 15:08:26.005434 4778 scope.go:117] "RemoveContainer" containerID="32e69f17a15da926e453ec4388e2482d274516709eb37f6124496feae6a6509f" Mar 12 15:08:28 crc kubenswrapper[4778]: I0312 15:08:28.557793 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:08:28 crc kubenswrapper[4778]: I0312 15:08:28.558386 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:08:28 crc kubenswrapper[4778]: I0312 15:08:28.558440 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 15:08:28 crc kubenswrapper[4778]: I0312 15:08:28.559240 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 15:08:28 crc kubenswrapper[4778]: I0312 15:08:28.559298 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" gracePeriod=600 Mar 12 15:08:28 crc kubenswrapper[4778]: E0312 15:08:28.695040 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:08:29 crc kubenswrapper[4778]: I0312 15:08:29.353265 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" exitCode=0 Mar 12 15:08:29 crc kubenswrapper[4778]: I0312 15:08:29.353327 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930"} Mar 12 15:08:29 crc kubenswrapper[4778]: I0312 15:08:29.353366 4778 scope.go:117] "RemoveContainer" containerID="009e612c3693545ba4a1988aa00993d05612427ec6eb485b08b455b35968f1ab" Mar 12 15:08:29 crc kubenswrapper[4778]: I0312 15:08:29.354135 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:08:29 crc kubenswrapper[4778]: E0312 15:08:29.354569 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:08:41 crc kubenswrapper[4778]: I0312 15:08:41.253895 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:08:41 crc kubenswrapper[4778]: E0312 15:08:41.254601 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:08:53 crc kubenswrapper[4778]: I0312 15:08:53.254221 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:08:53 crc kubenswrapper[4778]: E0312 15:08:53.254832 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:09:05 crc kubenswrapper[4778]: I0312 15:09:05.254480 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:09:05 crc kubenswrapper[4778]: E0312 15:09:05.255175 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:09:12 crc kubenswrapper[4778]: I0312 15:09:12.799107 4778 generic.go:334] "Generic (PLEG): container finished" podID="dd2baa0b-6680-41af-8231-e30368cb0090" containerID="099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb" exitCode=0 Mar 12 15:09:12 crc kubenswrapper[4778]: I0312 15:09:12.799174 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" event={"ID":"dd2baa0b-6680-41af-8231-e30368cb0090","Type":"ContainerDied","Data":"099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb"} Mar 12 15:09:12 crc kubenswrapper[4778]: I0312 15:09:12.800326 4778 scope.go:117] "RemoveContainer" containerID="099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb" Mar 12 15:09:13 crc kubenswrapper[4778]: I0312 15:09:13.751703 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rkpvq_must-gather-6d9ls_dd2baa0b-6680-41af-8231-e30368cb0090/gather/0.log" Mar 12 15:09:20 crc kubenswrapper[4778]: I0312 15:09:20.254734 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:09:20 crc kubenswrapper[4778]: E0312 15:09:20.255626 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.239094 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rkpvq/must-gather-6d9ls"] Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.239675 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="copy" containerID="cri-o://d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b" gracePeriod=2 Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.257640 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rkpvq/must-gather-6d9ls"] Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.715857 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rkpvq_must-gather-6d9ls_dd2baa0b-6680-41af-8231-e30368cb0090/copy/0.log" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.716704 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.806681 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dd2baa0b-6680-41af-8231-e30368cb0090-must-gather-output\") pod \"dd2baa0b-6680-41af-8231-e30368cb0090\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.806761 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twzlp\" (UniqueName: \"kubernetes.io/projected/dd2baa0b-6680-41af-8231-e30368cb0090-kube-api-access-twzlp\") pod \"dd2baa0b-6680-41af-8231-e30368cb0090\" (UID: \"dd2baa0b-6680-41af-8231-e30368cb0090\") " Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.813132 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2baa0b-6680-41af-8231-e30368cb0090-kube-api-access-twzlp" (OuterVolumeSpecName: "kube-api-access-twzlp") pod "dd2baa0b-6680-41af-8231-e30368cb0090" (UID: "dd2baa0b-6680-41af-8231-e30368cb0090"). InnerVolumeSpecName "kube-api-access-twzlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.909665 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twzlp\" (UniqueName: \"kubernetes.io/projected/dd2baa0b-6680-41af-8231-e30368cb0090-kube-api-access-twzlp\") on node \"crc\" DevicePath \"\"" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.916844 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rkpvq_must-gather-6d9ls_dd2baa0b-6680-41af-8231-e30368cb0090/copy/0.log" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.917161 4778 generic.go:334] "Generic (PLEG): container finished" podID="dd2baa0b-6680-41af-8231-e30368cb0090" containerID="d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b" exitCode=143 Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.917233 4778 scope.go:117] "RemoveContainer" containerID="d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b" Mar 12 15:09:23 crc kubenswrapper[4778]: I0312 15:09:23.917370 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rkpvq/must-gather-6d9ls" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.052960 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2baa0b-6680-41af-8231-e30368cb0090-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "dd2baa0b-6680-41af-8231-e30368cb0090" (UID: "dd2baa0b-6680-41af-8231-e30368cb0090"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.471370 4778 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dd2baa0b-6680-41af-8231-e30368cb0090-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.488829 4778 scope.go:117] "RemoveContainer" containerID="099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.491576 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" path="/var/lib/kubelet/pods/dd2baa0b-6680-41af-8231-e30368cb0090/volumes" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.579797 4778 scope.go:117] "RemoveContainer" containerID="d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b" Mar 12 15:09:24 crc kubenswrapper[4778]: E0312 15:09:24.581646 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b\": container with ID starting with d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b not found: ID does not exist" containerID="d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.581706 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b"} err="failed to get container status \"d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b\": rpc error: code = NotFound desc = could not find container \"d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b\": container with ID starting with d75c39511d2814b29f7a8f3d56db17a77d40e26925d76443f57f610bafcb652b not found: ID does not exist" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.581734 4778 scope.go:117] "RemoveContainer" containerID="099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb" Mar 12 15:09:24 crc kubenswrapper[4778]: E0312 15:09:24.582083 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb\": container with ID starting with 099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb not found: ID does not exist" containerID="099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb" Mar 12 15:09:24 crc kubenswrapper[4778]: I0312 15:09:24.582109 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb"} err="failed to get container status \"099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb\": rpc error: code = NotFound desc = could not find container \"099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb\": container with ID starting with 099ea78abdd9e205689d8ecdedd4eb5e53feb9c31e850ed5f759eb9dcba848eb not found: ID does not exist" Mar 12 15:09:26 crc kubenswrapper[4778]: I0312 15:09:26.131126 4778 scope.go:117] "RemoveContainer" containerID="74340f2038c644c2a2c001699df4f77fd8e1cf73ce4885bad06a1749c4f74a6f" Mar 12 15:09:26 crc kubenswrapper[4778]: I0312 15:09:26.151466 4778 scope.go:117] "RemoveContainer" containerID="e0b26a87b52c43c4e608ab6f59fc6362851c87a5e28dc029fccea83b1c3e5e7d" Mar 12 15:09:33 crc kubenswrapper[4778]: I0312 15:09:33.254446 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:09:33 crc kubenswrapper[4778]: E0312 15:09:33.255543 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:09:45 crc kubenswrapper[4778]: I0312 15:09:45.254536 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:09:45 crc kubenswrapper[4778]: E0312 15:09:45.257214 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:09:59 crc kubenswrapper[4778]: I0312 15:09:59.254655 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:09:59 crc kubenswrapper[4778]: E0312 15:09:59.255410 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.155326 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555470-65l68"] Mar 12 15:10:00 crc kubenswrapper[4778]: E0312 15:10:00.156081 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="gather" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156103 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="gather" Mar 12 15:10:00 crc kubenswrapper[4778]: E0312 15:10:00.156128 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="copy" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156134 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="copy" Mar 12 15:10:00 crc kubenswrapper[4778]: E0312 15:10:00.156147 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="extract-content" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156154 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="extract-content" Mar 12 15:10:00 crc kubenswrapper[4778]: E0312 15:10:00.156164 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="extract-utilities" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156171 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="extract-utilities" Mar 12 15:10:00 crc kubenswrapper[4778]: E0312 15:10:00.156197 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="registry-server" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156203 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="registry-server" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156406 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="gather" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156417 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2baa0b-6680-41af-8231-e30368cb0090" containerName="copy" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.156432 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="193a1938-83bf-48dc-abd9-ecc2f202db8d" containerName="registry-server" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.157042 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.160525 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.160880 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.161450 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.171487 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555470-65l68"] Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.270253 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4lng\" (UniqueName: \"kubernetes.io/projected/f05c65b6-74a0-49ef-8f84-3b4453313dc7-kube-api-access-b4lng\") pod \"auto-csr-approver-29555470-65l68\" (UID: \"f05c65b6-74a0-49ef-8f84-3b4453313dc7\") " pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.372483 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4lng\" (UniqueName: \"kubernetes.io/projected/f05c65b6-74a0-49ef-8f84-3b4453313dc7-kube-api-access-b4lng\") pod \"auto-csr-approver-29555470-65l68\" (UID: \"f05c65b6-74a0-49ef-8f84-3b4453313dc7\") " pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.391784 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4lng\" (UniqueName: \"kubernetes.io/projected/f05c65b6-74a0-49ef-8f84-3b4453313dc7-kube-api-access-b4lng\") pod \"auto-csr-approver-29555470-65l68\" (UID: \"f05c65b6-74a0-49ef-8f84-3b4453313dc7\") " pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.487585 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.936714 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555470-65l68"] Mar 12 15:10:00 crc kubenswrapper[4778]: I0312 15:10:00.940381 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 15:10:01 crc kubenswrapper[4778]: I0312 15:10:01.283647 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555470-65l68" event={"ID":"f05c65b6-74a0-49ef-8f84-3b4453313dc7","Type":"ContainerStarted","Data":"6da4a4d0c09739dff2ab3fa082e65fb673da228e6d94f3fd22b17604e2105a4a"} Mar 12 15:10:03 crc kubenswrapper[4778]: I0312 15:10:03.306954 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555470-65l68" event={"ID":"f05c65b6-74a0-49ef-8f84-3b4453313dc7","Type":"ContainerStarted","Data":"66d173277dbb8cde37f4f992e677953055661368f74064cf032011267c61214c"} Mar 12 15:10:03 crc kubenswrapper[4778]: I0312 15:10:03.326375 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555470-65l68" podStartSLOduration=1.6866532969999999 podStartE2EDuration="3.326353807s" podCreationTimestamp="2026-03-12 15:10:00 +0000 UTC" firstStartedPulling="2026-03-12 15:10:00.940100812 +0000 UTC m=+7219.388796208" lastFinishedPulling="2026-03-12 15:10:02.579801312 +0000 UTC m=+7221.028496718" observedRunningTime="2026-03-12 15:10:03.320698936 +0000 UTC m=+7221.769394342" watchObservedRunningTime="2026-03-12 15:10:03.326353807 +0000 UTC m=+7221.775049203" Mar 12 15:10:04 crc kubenswrapper[4778]: I0312 15:10:04.316405 4778 generic.go:334] "Generic (PLEG): container finished" podID="f05c65b6-74a0-49ef-8f84-3b4453313dc7" containerID="66d173277dbb8cde37f4f992e677953055661368f74064cf032011267c61214c" exitCode=0 Mar 12 15:10:04 crc kubenswrapper[4778]: I0312 15:10:04.316716 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555470-65l68" event={"ID":"f05c65b6-74a0-49ef-8f84-3b4453313dc7","Type":"ContainerDied","Data":"66d173277dbb8cde37f4f992e677953055661368f74064cf032011267c61214c"} Mar 12 15:10:05 crc kubenswrapper[4778]: I0312 15:10:05.705894 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:05 crc kubenswrapper[4778]: I0312 15:10:05.774569 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4lng\" (UniqueName: \"kubernetes.io/projected/f05c65b6-74a0-49ef-8f84-3b4453313dc7-kube-api-access-b4lng\") pod \"f05c65b6-74a0-49ef-8f84-3b4453313dc7\" (UID: \"f05c65b6-74a0-49ef-8f84-3b4453313dc7\") " Mar 12 15:10:05 crc kubenswrapper[4778]: I0312 15:10:05.782531 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05c65b6-74a0-49ef-8f84-3b4453313dc7-kube-api-access-b4lng" (OuterVolumeSpecName: "kube-api-access-b4lng") pod "f05c65b6-74a0-49ef-8f84-3b4453313dc7" (UID: "f05c65b6-74a0-49ef-8f84-3b4453313dc7"). InnerVolumeSpecName "kube-api-access-b4lng". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:10:05 crc kubenswrapper[4778]: I0312 15:10:05.877672 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4lng\" (UniqueName: \"kubernetes.io/projected/f05c65b6-74a0-49ef-8f84-3b4453313dc7-kube-api-access-b4lng\") on node \"crc\" DevicePath \"\"" Mar 12 15:10:06 crc kubenswrapper[4778]: I0312 15:10:06.338535 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555470-65l68" event={"ID":"f05c65b6-74a0-49ef-8f84-3b4453313dc7","Type":"ContainerDied","Data":"6da4a4d0c09739dff2ab3fa082e65fb673da228e6d94f3fd22b17604e2105a4a"} Mar 12 15:10:06 crc kubenswrapper[4778]: I0312 15:10:06.338950 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6da4a4d0c09739dff2ab3fa082e65fb673da228e6d94f3fd22b17604e2105a4a" Mar 12 15:10:06 crc kubenswrapper[4778]: I0312 15:10:06.338781 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555470-65l68" Mar 12 15:10:06 crc kubenswrapper[4778]: I0312 15:10:06.409729 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555464-rt6fz"] Mar 12 15:10:06 crc kubenswrapper[4778]: I0312 15:10:06.417672 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555464-rt6fz"] Mar 12 15:10:08 crc kubenswrapper[4778]: I0312 15:10:08.264362 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc35b0a-1b16-4db8-adef-8a6afd6ae934" path="/var/lib/kubelet/pods/bdc35b0a-1b16-4db8-adef-8a6afd6ae934/volumes" Mar 12 15:10:11 crc kubenswrapper[4778]: I0312 15:10:11.254116 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:10:11 crc kubenswrapper[4778]: E0312 15:10:11.255902 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:10:25 crc kubenswrapper[4778]: I0312 15:10:25.258627 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:10:25 crc kubenswrapper[4778]: E0312 15:10:25.259871 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:10:26 crc kubenswrapper[4778]: I0312 15:10:26.270121 4778 scope.go:117] "RemoveContainer" containerID="13a5daaa89f4db1da5ea953e47f7efe223f79130e65c2950cad2f908b81e7d1b" Mar 12 15:10:37 crc kubenswrapper[4778]: I0312 15:10:37.254162 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:10:37 crc kubenswrapper[4778]: E0312 15:10:37.255099 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:10:48 crc kubenswrapper[4778]: I0312 15:10:48.254305 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:10:48 crc kubenswrapper[4778]: E0312 15:10:48.255126 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:11:01 crc kubenswrapper[4778]: I0312 15:11:01.255092 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:11:01 crc kubenswrapper[4778]: E0312 15:11:01.256149 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:11:16 crc kubenswrapper[4778]: I0312 15:11:16.254342 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:11:16 crc kubenswrapper[4778]: E0312 15:11:16.255245 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:11:31 crc kubenswrapper[4778]: I0312 15:11:31.254888 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:11:31 crc kubenswrapper[4778]: E0312 15:11:31.256170 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:11:46 crc kubenswrapper[4778]: I0312 15:11:46.253461 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:11:46 crc kubenswrapper[4778]: E0312 15:11:46.254269 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:11:58 crc kubenswrapper[4778]: I0312 15:11:58.253988 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:11:58 crc kubenswrapper[4778]: E0312 15:11:58.254821 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.169251 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555472-k4f9h"] Mar 12 15:12:00 crc kubenswrapper[4778]: E0312 15:12:00.170117 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05c65b6-74a0-49ef-8f84-3b4453313dc7" containerName="oc" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.170132 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05c65b6-74a0-49ef-8f84-3b4453313dc7" containerName="oc" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.170396 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f05c65b6-74a0-49ef-8f84-3b4453313dc7" containerName="oc" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.171261 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.174117 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.174561 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.175140 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.189341 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555472-k4f9h"] Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.344049 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5mxz\" (UniqueName: \"kubernetes.io/projected/30c4e913-d163-4764-8738-ac336cd93df9-kube-api-access-x5mxz\") pod \"auto-csr-approver-29555472-k4f9h\" (UID: \"30c4e913-d163-4764-8738-ac336cd93df9\") " pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.446161 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5mxz\" (UniqueName: \"kubernetes.io/projected/30c4e913-d163-4764-8738-ac336cd93df9-kube-api-access-x5mxz\") pod \"auto-csr-approver-29555472-k4f9h\" (UID: \"30c4e913-d163-4764-8738-ac336cd93df9\") " pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.465498 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5mxz\" (UniqueName: \"kubernetes.io/projected/30c4e913-d163-4764-8738-ac336cd93df9-kube-api-access-x5mxz\") pod \"auto-csr-approver-29555472-k4f9h\" (UID: \"30c4e913-d163-4764-8738-ac336cd93df9\") " pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.502895 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:00 crc kubenswrapper[4778]: I0312 15:12:00.965958 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555472-k4f9h"] Mar 12 15:12:00 crc kubenswrapper[4778]: W0312 15:12:00.966899 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30c4e913_d163_4764_8738_ac336cd93df9.slice/crio-24ce58de7517e21c1f2011e5b11909f2e95fc03f93ab0a4c71d72ac7864c9440 WatchSource:0}: Error finding container 24ce58de7517e21c1f2011e5b11909f2e95fc03f93ab0a4c71d72ac7864c9440: Status 404 returned error can't find the container with id 24ce58de7517e21c1f2011e5b11909f2e95fc03f93ab0a4c71d72ac7864c9440 Mar 12 15:12:01 crc kubenswrapper[4778]: I0312 15:12:01.595318 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" event={"ID":"30c4e913-d163-4764-8738-ac336cd93df9","Type":"ContainerStarted","Data":"24ce58de7517e21c1f2011e5b11909f2e95fc03f93ab0a4c71d72ac7864c9440"} Mar 12 15:12:02 crc kubenswrapper[4778]: I0312 15:12:02.603425 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" event={"ID":"30c4e913-d163-4764-8738-ac336cd93df9","Type":"ContainerStarted","Data":"971c448e63690dd43ac1d65335a70f73b2547d4337b42531c9336354c82b33f3"} Mar 12 15:12:02 crc kubenswrapper[4778]: I0312 15:12:02.622195 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" podStartSLOduration=1.560282964 podStartE2EDuration="2.622156819s" podCreationTimestamp="2026-03-12 15:12:00 +0000 UTC" firstStartedPulling="2026-03-12 15:12:00.969292154 +0000 UTC m=+7339.417987570" lastFinishedPulling="2026-03-12 15:12:02.031166029 +0000 UTC m=+7340.479861425" observedRunningTime="2026-03-12 15:12:02.614687857 +0000 UTC m=+7341.063383253" watchObservedRunningTime="2026-03-12 15:12:02.622156819 +0000 UTC m=+7341.070852205" Mar 12 15:12:03 crc kubenswrapper[4778]: I0312 15:12:03.614619 4778 generic.go:334] "Generic (PLEG): container finished" podID="30c4e913-d163-4764-8738-ac336cd93df9" containerID="971c448e63690dd43ac1d65335a70f73b2547d4337b42531c9336354c82b33f3" exitCode=0 Mar 12 15:12:03 crc kubenswrapper[4778]: I0312 15:12:03.614659 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" event={"ID":"30c4e913-d163-4764-8738-ac336cd93df9","Type":"ContainerDied","Data":"971c448e63690dd43ac1d65335a70f73b2547d4337b42531c9336354c82b33f3"} Mar 12 15:12:04 crc kubenswrapper[4778]: I0312 15:12:04.981630 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.065045 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5mxz\" (UniqueName: \"kubernetes.io/projected/30c4e913-d163-4764-8738-ac336cd93df9-kube-api-access-x5mxz\") pod \"30c4e913-d163-4764-8738-ac336cd93df9\" (UID: \"30c4e913-d163-4764-8738-ac336cd93df9\") " Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.071596 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c4e913-d163-4764-8738-ac336cd93df9-kube-api-access-x5mxz" (OuterVolumeSpecName: "kube-api-access-x5mxz") pod "30c4e913-d163-4764-8738-ac336cd93df9" (UID: "30c4e913-d163-4764-8738-ac336cd93df9"). InnerVolumeSpecName "kube-api-access-x5mxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.167003 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5mxz\" (UniqueName: \"kubernetes.io/projected/30c4e913-d163-4764-8738-ac336cd93df9-kube-api-access-x5mxz\") on node \"crc\" DevicePath \"\"" Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.369002 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555466-7wlhj"] Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.376929 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555466-7wlhj"] Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.645667 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" event={"ID":"30c4e913-d163-4764-8738-ac336cd93df9","Type":"ContainerDied","Data":"24ce58de7517e21c1f2011e5b11909f2e95fc03f93ab0a4c71d72ac7864c9440"} Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.645726 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24ce58de7517e21c1f2011e5b11909f2e95fc03f93ab0a4c71d72ac7864c9440" Mar 12 15:12:05 crc kubenswrapper[4778]: I0312 15:12:05.645743 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555472-k4f9h" Mar 12 15:12:06 crc kubenswrapper[4778]: I0312 15:12:06.268495 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58e0f99-4ece-49ec-9c47-b82055df7d48" path="/var/lib/kubelet/pods/c58e0f99-4ece-49ec-9c47-b82055df7d48/volumes" Mar 12 15:12:12 crc kubenswrapper[4778]: I0312 15:12:12.261281 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:12:12 crc kubenswrapper[4778]: E0312 15:12:12.262090 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:12:24 crc kubenswrapper[4778]: I0312 15:12:24.611059 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-86cb765474-5pq5z_6bd172c5-383f-4273-98a5-2c92223dc765/barbican-api-log/0.log" Mar 12 15:12:25 crc kubenswrapper[4778]: I0312 15:12:25.448114 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65c9994dfd-xznqh_8ee1f546-8428-4b23-93e4-b8370fd4224b/barbican-keystone-listener-log/0.log" Mar 12 15:12:25 crc kubenswrapper[4778]: I0312 15:12:25.952898 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7dcf9787-ngc87_d505bb59-3c9e-4cfa-891c-c8e0068e2567/barbican-worker-log/0.log" Mar 12 15:12:26 crc kubenswrapper[4778]: I0312 15:12:26.365449 4778 scope.go:117] "RemoveContainer" containerID="8984f879d02ecea61666d68e6857174dd681c26238a5c27d5a617cc7dccda3db" Mar 12 15:12:26 crc kubenswrapper[4778]: I0312 15:12:26.527899 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx_b99627a8-43d8-4f7d-90f7-530eda3c2213/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:27 crc kubenswrapper[4778]: I0312 15:12:27.077328 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/ceilometer-central-agent/0.log" Mar 12 15:12:27 crc kubenswrapper[4778]: I0312 15:12:27.254456 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:12:27 crc kubenswrapper[4778]: E0312 15:12:27.254681 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:12:27 crc kubenswrapper[4778]: I0312 15:12:27.598703 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_99f72014-50e8-4dd4-9764-1b2c7d546b30/cinder-api-log/0.log" Mar 12 15:12:28 crc kubenswrapper[4778]: I0312 15:12:28.116448 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_39ee2404-53a8-4598-8c4b-c3a34fbf3480/cinder-scheduler/0.log" Mar 12 15:12:28 crc kubenswrapper[4778]: I0312 15:12:28.613323 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4szjl_5c5541f3-fb44-476b-91c2-b07dffe50894/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:29 crc kubenswrapper[4778]: I0312 15:12:29.108687 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6_36bb4acd-fab3-4998-a8cd-a6ebcc800fc8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:30 crc kubenswrapper[4778]: I0312 15:12:30.212543 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/dnsmasq-dns/0.log" Mar 12 15:12:30 crc kubenswrapper[4778]: I0312 15:12:30.767178 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-2xksx_96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:31 crc kubenswrapper[4778]: I0312 15:12:31.275339 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_81c1a05c-5642-43d4-8a7b-229330168332/glance-log/0.log" Mar 12 15:12:31 crc kubenswrapper[4778]: I0312 15:12:31.792408 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7fa757af-1c91-4b93-8916-5bbd99b8522e/glance-log/0.log" Mar 12 15:12:32 crc kubenswrapper[4778]: I0312 15:12:32.332406 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bngcx_f69e6cfe-f7c2-4127-b4df-710725c52227/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:32 crc kubenswrapper[4778]: I0312 15:12:32.808649 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g252n_29f8609b-4a3b-42ba-9450-a2b633bb4c2c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:34 crc kubenswrapper[4778]: I0312 15:12:34.625971 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-6lrlq_a56bb599-f10d-4564-b6bf-48128dc2c7f1/keystone-api/0.log" Mar 12 15:12:36 crc kubenswrapper[4778]: I0312 15:12:36.502206 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-z4h9m_16dea17b-eaa4-4bbf-8895-c077b3e28d66/keystone-api/0.log" Mar 12 15:12:36 crc kubenswrapper[4778]: I0312 15:12:36.961492 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555401-vjgkl_e4df6927-3452-4b36-b59a-a1fdcd4272a4/keystone-cron/0.log" Mar 12 15:12:37 crc kubenswrapper[4778]: I0312 15:12:37.402972 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555461-lmqk9_ebdf3274-70cb-4083-bf12-5d1038a9b7ba/keystone-cron/0.log" Mar 12 15:12:37 crc kubenswrapper[4778]: I0312 15:12:37.899456 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_51f24fcd-aff5-4785-abf7-4936180cee78/kube-state-metrics/0.log" Mar 12 15:12:38 crc kubenswrapper[4778]: I0312 15:12:38.501424 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8_8713b951-b516-42bd-9286-4343e5bcc955/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:39 crc kubenswrapper[4778]: I0312 15:12:39.253865 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:12:39 crc kubenswrapper[4778]: E0312 15:12:39.254476 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:12:39 crc kubenswrapper[4778]: I0312 15:12:39.863764 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ec63cc68-6fde-419b-973c-91fc982e6a49/memcached/0.log" Mar 12 15:12:44 crc kubenswrapper[4778]: I0312 15:12:44.227049 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-dggmh_7596a69e-33c9-4a2b-89fc-e4c41252b3fd/neutron-api/0.log" Mar 12 15:12:47 crc kubenswrapper[4778]: I0312 15:12:47.955498 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-zx97x_8a67d4b7-d8eb-40f4-b51d-62e92c6042c1/neutron-api/0.log" Mar 12 15:12:48 crc kubenswrapper[4778]: I0312 15:12:48.431361 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg_5cc410de-5b42-44d1-8b29-37161475730e/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:50 crc kubenswrapper[4778]: I0312 15:12:50.434000 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4/nova-api-log/0.log" Mar 12 15:12:52 crc kubenswrapper[4778]: I0312 15:12:52.256018 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-1_f0341d80-4327-4c9e-bc11-0cddbc6eab66/nova-api-log/0.log" Mar 12 15:12:53 crc kubenswrapper[4778]: I0312 15:12:53.262681 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_929bb450-949d-4f4f-9c21-de6c3fe32927/nova-cell0-conductor-conductor/0.log" Mar 12 15:12:53 crc kubenswrapper[4778]: I0312 15:12:53.898890 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1466aea3-fa10-49a6-a254-a96a52091aca/nova-cell1-conductor-conductor/0.log" Mar 12 15:12:54 crc kubenswrapper[4778]: I0312 15:12:54.255108 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:12:54 crc kubenswrapper[4778]: E0312 15:12:54.255736 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:12:54 crc kubenswrapper[4778]: I0312 15:12:54.500967 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-metadata-0_c289a520-78eb-433f-b7a4-0c03be917c18/nova-cell1-metadata-log/0.log" Mar 12 15:12:55 crc kubenswrapper[4778]: I0312 15:12:55.083228 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7/nova-cell1-novncproxy-novncproxy/0.log" Mar 12 15:12:55 crc kubenswrapper[4778]: I0312 15:12:55.683730 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5tw6s_6ed77f87-e6b2-4c7a-8b0e-003106200dc8/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:12:57 crc kubenswrapper[4778]: I0312 15:12:57.229437 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f613745b-fe33-4918-9e0a-da2a59c55e33/nova-scheduler-scheduler/0.log" Mar 12 15:12:57 crc kubenswrapper[4778]: I0312 15:12:57.747139 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/galera/0.log" Mar 12 15:12:58 crc kubenswrapper[4778]: I0312 15:12:58.216427 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/galera/0.log" Mar 12 15:12:58 crc kubenswrapper[4778]: I0312 15:12:58.670368 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_856cd6d1-db21-4503-94d7-cbf27ca96cc2/openstackclient/0.log" Mar 12 15:12:59 crc kubenswrapper[4778]: I0312 15:12:59.155449 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4wct6_3b8efd1e-884d-4963-b69f-04ede0a92267/ovn-controller/0.log" Mar 12 15:12:59 crc kubenswrapper[4778]: I0312 15:12:59.563865 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vtt4z_a8484e5d-6f77-407c-81db-0d9b2a6b37fd/openstack-network-exporter/0.log" Mar 12 15:13:00 crc kubenswrapper[4778]: I0312 15:13:00.041281 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server/0.log" Mar 12 15:13:00 crc kubenswrapper[4778]: I0312 15:13:00.601032 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9lbdq_3c0a2200-506d-4ac3-b08c-9b3156c9e573/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:13:01 crc kubenswrapper[4778]: I0312 15:13:01.548222 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b25f9c9-784a-4a52-9bb3-02c6c4592702/ovn-northd/0.log" Mar 12 15:13:02 crc kubenswrapper[4778]: I0312 15:13:02.077586 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7321e15e-673c-4e0d-80f8-6ac644c1940f/ovsdbserver-nb/0.log" Mar 12 15:13:02 crc kubenswrapper[4778]: I0312 15:13:02.591575 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c951c6f-06fd-4793-a95b-26b5c1400d73/ovsdbserver-sb/0.log" Mar 12 15:13:03 crc kubenswrapper[4778]: I0312 15:13:03.613588 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d4d765698-l7bjx_267e7df2-d35c-45c4-af65-e8af31f8f6cf/placement-log/0.log" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.270686 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/rabbitmq/0.log" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.774920 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7cktf"] Mar 12 15:13:04 crc kubenswrapper[4778]: E0312 15:13:04.775339 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c4e913-d163-4764-8738-ac336cd93df9" containerName="oc" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.775351 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c4e913-d163-4764-8738-ac336cd93df9" containerName="oc" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.775561 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c4e913-d163-4764-8738-ac336cd93df9" containerName="oc" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.777154 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.869945 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7cktf"] Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.889342 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-catalog-content\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.889433 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzsf2\" (UniqueName: \"kubernetes.io/projected/690208cb-cbdb-488c-9998-70cf01f1cc05-kube-api-access-dzsf2\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.889469 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-utilities\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.911737 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/rabbitmq/0.log" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.991239 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-catalog-content\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.991296 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzsf2\" (UniqueName: \"kubernetes.io/projected/690208cb-cbdb-488c-9998-70cf01f1cc05-kube-api-access-dzsf2\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.991317 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-utilities\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.991776 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-utilities\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:04 crc kubenswrapper[4778]: I0312 15:13:04.992018 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-catalog-content\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:05 crc kubenswrapper[4778]: I0312 15:13:05.011803 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzsf2\" (UniqueName: \"kubernetes.io/projected/690208cb-cbdb-488c-9998-70cf01f1cc05-kube-api-access-dzsf2\") pod \"community-operators-7cktf\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:05 crc kubenswrapper[4778]: I0312 15:13:05.159656 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:05 crc kubenswrapper[4778]: I0312 15:13:05.449406 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc_43a3ffe4-8b64-4e26-b63a-5254a986e4a4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:13:05 crc kubenswrapper[4778]: I0312 15:13:05.702748 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7cktf"] Mar 12 15:13:05 crc kubenswrapper[4778]: W0312 15:13:05.716543 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod690208cb_cbdb_488c_9998_70cf01f1cc05.slice/crio-b4c4d370a37ff718b518066aa455a41a074b334ef4b1257119931ced115773bd WatchSource:0}: Error finding container b4c4d370a37ff718b518066aa455a41a074b334ef4b1257119931ced115773bd: Status 404 returned error can't find the container with id b4c4d370a37ff718b518066aa455a41a074b334ef4b1257119931ced115773bd Mar 12 15:13:05 crc kubenswrapper[4778]: I0312 15:13:05.886212 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc_bd7ac6b4-5600-45ce-b0ea-199dd4baefcb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:13:06 crc kubenswrapper[4778]: I0312 15:13:06.253811 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:13:06 crc kubenswrapper[4778]: E0312 15:13:06.254414 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:13:06 crc kubenswrapper[4778]: I0312 15:13:06.272903 4778 generic.go:334] "Generic (PLEG): container finished" podID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerID="e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7" exitCode=0 Mar 12 15:13:06 crc kubenswrapper[4778]: I0312 15:13:06.272952 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerDied","Data":"e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7"} Mar 12 15:13:06 crc kubenswrapper[4778]: I0312 15:13:06.272982 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerStarted","Data":"b4c4d370a37ff718b518066aa455a41a074b334ef4b1257119931ced115773bd"} Mar 12 15:13:06 crc kubenswrapper[4778]: I0312 15:13:06.333314 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gt58t_b0bb06df-44bb-4939-9492-a6ad3d6b5368/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:13:06 crc kubenswrapper[4778]: I0312 15:13:06.905130 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8mmjm_c993b33e-6c36-4524-864a-65da461a8e0c/ssh-known-hosts-edpm-deployment/0.log" Mar 12 15:13:07 crc kubenswrapper[4778]: I0312 15:13:07.284837 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerStarted","Data":"07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca"} Mar 12 15:13:07 crc kubenswrapper[4778]: I0312 15:13:07.744403 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-77f887c49f-fw2qd_bbd76cb8-462f-4e60-b755-ef3170e70d11/proxy-httpd/0.log" Mar 12 15:13:08 crc kubenswrapper[4778]: I0312 15:13:08.283959 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5knbg_2edc2c90-f91e-402d-809c-514e9d8a5e04/swift-ring-rebalance/0.log" Mar 12 15:13:08 crc kubenswrapper[4778]: I0312 15:13:08.297456 4778 generic.go:334] "Generic (PLEG): container finished" podID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerID="07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca" exitCode=0 Mar 12 15:13:08 crc kubenswrapper[4778]: I0312 15:13:08.297519 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerDied","Data":"07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca"} Mar 12 15:13:08 crc kubenswrapper[4778]: I0312 15:13:08.861294 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-server/0.log" Mar 12 15:13:09 crc kubenswrapper[4778]: I0312 15:13:09.308746 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerStarted","Data":"cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f"} Mar 12 15:13:09 crc kubenswrapper[4778]: I0312 15:13:09.338871 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7cktf" podStartSLOduration=2.849708025 podStartE2EDuration="5.338849147s" podCreationTimestamp="2026-03-12 15:13:04 +0000 UTC" firstStartedPulling="2026-03-12 15:13:06.274510024 +0000 UTC m=+7404.723205420" lastFinishedPulling="2026-03-12 15:13:08.763651146 +0000 UTC m=+7407.212346542" observedRunningTime="2026-03-12 15:13:09.326806405 +0000 UTC m=+7407.775501801" watchObservedRunningTime="2026-03-12 15:13:09.338849147 +0000 UTC m=+7407.787544543" Mar 12 15:13:09 crc kubenswrapper[4778]: I0312 15:13:09.404756 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s_2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:13:09 crc kubenswrapper[4778]: I0312 15:13:09.892545 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_74897d0a-ca7b-4589-bd4c-75910c2d491c/tempest-tests-tempest-tests-runner/0.log" Mar 12 15:13:10 crc kubenswrapper[4778]: I0312 15:13:10.322510 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_82246f69-2112-44e9-a783-a4a5926188b4/test-operator-logs-container/0.log" Mar 12 15:13:10 crc kubenswrapper[4778]: I0312 15:13:10.801378 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9glvr_41583476-38cd-4c0d-a05a-96ddc5b330ca/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.154147 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fdrpx"] Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.157208 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.194125 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdrpx"] Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.235947 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrzgn\" (UniqueName: \"kubernetes.io/projected/8549fb56-bc75-4c66-8900-ba62a687ce0e-kube-api-access-nrzgn\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.236003 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-catalog-content\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.236349 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-utilities\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.338715 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrzgn\" (UniqueName: \"kubernetes.io/projected/8549fb56-bc75-4c66-8900-ba62a687ce0e-kube-api-access-nrzgn\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.339132 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-catalog-content\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.339555 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-utilities\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.339874 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-catalog-content\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.340010 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-utilities\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.369389 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrzgn\" (UniqueName: \"kubernetes.io/projected/8549fb56-bc75-4c66-8900-ba62a687ce0e-kube-api-access-nrzgn\") pod \"redhat-marketplace-fdrpx\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:12 crc kubenswrapper[4778]: I0312 15:13:12.483122 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:13 crc kubenswrapper[4778]: I0312 15:13:13.059716 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdrpx"] Mar 12 15:13:13 crc kubenswrapper[4778]: I0312 15:13:13.346888 4778 generic.go:334] "Generic (PLEG): container finished" podID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerID="9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8" exitCode=0 Mar 12 15:13:13 crc kubenswrapper[4778]: I0312 15:13:13.346953 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerDied","Data":"9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8"} Mar 12 15:13:13 crc kubenswrapper[4778]: I0312 15:13:13.347016 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerStarted","Data":"74af807183905ac63573678f5f2eb3366ccf037a22980bf197b990c23186ec82"} Mar 12 15:13:14 crc kubenswrapper[4778]: I0312 15:13:14.360368 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerStarted","Data":"808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762"} Mar 12 15:13:15 crc kubenswrapper[4778]: I0312 15:13:15.160816 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:15 crc kubenswrapper[4778]: I0312 15:13:15.161255 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:15 crc kubenswrapper[4778]: I0312 15:13:15.206128 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:15 crc kubenswrapper[4778]: I0312 15:13:15.374111 4778 generic.go:334] "Generic (PLEG): container finished" podID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerID="808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762" exitCode=0 Mar 12 15:13:15 crc kubenswrapper[4778]: I0312 15:13:15.374219 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerDied","Data":"808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762"} Mar 12 15:13:15 crc kubenswrapper[4778]: I0312 15:13:15.424158 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:16 crc kubenswrapper[4778]: I0312 15:13:16.383963 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerStarted","Data":"fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3"} Mar 12 15:13:16 crc kubenswrapper[4778]: I0312 15:13:16.407301 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fdrpx" podStartSLOduration=1.712659197 podStartE2EDuration="4.407282373s" podCreationTimestamp="2026-03-12 15:13:12 +0000 UTC" firstStartedPulling="2026-03-12 15:13:13.349355223 +0000 UTC m=+7411.798050619" lastFinishedPulling="2026-03-12 15:13:16.043978389 +0000 UTC m=+7414.492673795" observedRunningTime="2026-03-12 15:13:16.401916111 +0000 UTC m=+7414.850611517" watchObservedRunningTime="2026-03-12 15:13:16.407282373 +0000 UTC m=+7414.855977769" Mar 12 15:13:17 crc kubenswrapper[4778]: I0312 15:13:17.538055 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7cktf"] Mar 12 15:13:17 crc kubenswrapper[4778]: I0312 15:13:17.539123 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7cktf" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="registry-server" containerID="cri-o://cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f" gracePeriod=2 Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.034686 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.096213 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzsf2\" (UniqueName: \"kubernetes.io/projected/690208cb-cbdb-488c-9998-70cf01f1cc05-kube-api-access-dzsf2\") pod \"690208cb-cbdb-488c-9998-70cf01f1cc05\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.096301 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-utilities\") pod \"690208cb-cbdb-488c-9998-70cf01f1cc05\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.096510 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-catalog-content\") pod \"690208cb-cbdb-488c-9998-70cf01f1cc05\" (UID: \"690208cb-cbdb-488c-9998-70cf01f1cc05\") " Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.098010 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-utilities" (OuterVolumeSpecName: "utilities") pod "690208cb-cbdb-488c-9998-70cf01f1cc05" (UID: "690208cb-cbdb-488c-9998-70cf01f1cc05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.119356 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690208cb-cbdb-488c-9998-70cf01f1cc05-kube-api-access-dzsf2" (OuterVolumeSpecName: "kube-api-access-dzsf2") pod "690208cb-cbdb-488c-9998-70cf01f1cc05" (UID: "690208cb-cbdb-488c-9998-70cf01f1cc05"). InnerVolumeSpecName "kube-api-access-dzsf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.171245 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "690208cb-cbdb-488c-9998-70cf01f1cc05" (UID: "690208cb-cbdb-488c-9998-70cf01f1cc05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.199485 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzsf2\" (UniqueName: \"kubernetes.io/projected/690208cb-cbdb-488c-9998-70cf01f1cc05-kube-api-access-dzsf2\") on node \"crc\" DevicePath \"\"" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.199531 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.199545 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690208cb-cbdb-488c-9998-70cf01f1cc05-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.403656 4778 generic.go:334] "Generic (PLEG): container finished" podID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerID="cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f" exitCode=0 Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.403702 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cktf" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.403709 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerDied","Data":"cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f"} Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.403741 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cktf" event={"ID":"690208cb-cbdb-488c-9998-70cf01f1cc05","Type":"ContainerDied","Data":"b4c4d370a37ff718b518066aa455a41a074b334ef4b1257119931ced115773bd"} Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.403762 4778 scope.go:117] "RemoveContainer" containerID="cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.431606 4778 scope.go:117] "RemoveContainer" containerID="07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.434131 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7cktf"] Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.445917 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7cktf"] Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.459054 4778 scope.go:117] "RemoveContainer" containerID="e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.521099 4778 scope.go:117] "RemoveContainer" containerID="cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f" Mar 12 15:13:18 crc kubenswrapper[4778]: E0312 15:13:18.522118 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f\": container with ID starting with cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f not found: ID does not exist" containerID="cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.522168 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f"} err="failed to get container status \"cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f\": rpc error: code = NotFound desc = could not find container \"cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f\": container with ID starting with cf84feed15251158f0773744cf0837c36cc843b6cf86c92e072869bcc5f3a48f not found: ID does not exist" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.522220 4778 scope.go:117] "RemoveContainer" containerID="07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca" Mar 12 15:13:18 crc kubenswrapper[4778]: E0312 15:13:18.522626 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca\": container with ID starting with 07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca not found: ID does not exist" containerID="07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.522664 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca"} err="failed to get container status \"07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca\": rpc error: code = NotFound desc = could not find container \"07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca\": container with ID starting with 07e287f5f2855697f874a35c1d2ba69a376cb020469f9688118410de38eee4ca not found: ID does not exist" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.522688 4778 scope.go:117] "RemoveContainer" containerID="e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7" Mar 12 15:13:18 crc kubenswrapper[4778]: E0312 15:13:18.523057 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7\": container with ID starting with e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7 not found: ID does not exist" containerID="e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7" Mar 12 15:13:18 crc kubenswrapper[4778]: I0312 15:13:18.523085 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7"} err="failed to get container status \"e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7\": rpc error: code = NotFound desc = could not find container \"e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7\": container with ID starting with e77193c08ece5fcdfb900c49597028aa2273d0551c0b231ebb63709f63e11ad7 not found: ID does not exist" Mar 12 15:13:20 crc kubenswrapper[4778]: I0312 15:13:20.270147 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" path="/var/lib/kubelet/pods/690208cb-cbdb-488c-9998-70cf01f1cc05/volumes" Mar 12 15:13:21 crc kubenswrapper[4778]: I0312 15:13:21.254964 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:13:21 crc kubenswrapper[4778]: E0312 15:13:21.255497 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:13:22 crc kubenswrapper[4778]: I0312 15:13:22.484017 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:22 crc kubenswrapper[4778]: I0312 15:13:22.484497 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:22 crc kubenswrapper[4778]: I0312 15:13:22.544596 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:23 crc kubenswrapper[4778]: I0312 15:13:23.541893 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:23 crc kubenswrapper[4778]: I0312 15:13:23.594497 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdrpx"] Mar 12 15:13:25 crc kubenswrapper[4778]: I0312 15:13:25.514275 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fdrpx" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="registry-server" containerID="cri-o://fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3" gracePeriod=2 Mar 12 15:13:25 crc kubenswrapper[4778]: I0312 15:13:25.984171 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.174353 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrzgn\" (UniqueName: \"kubernetes.io/projected/8549fb56-bc75-4c66-8900-ba62a687ce0e-kube-api-access-nrzgn\") pod \"8549fb56-bc75-4c66-8900-ba62a687ce0e\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.174594 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-utilities\") pod \"8549fb56-bc75-4c66-8900-ba62a687ce0e\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.174886 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-catalog-content\") pod \"8549fb56-bc75-4c66-8900-ba62a687ce0e\" (UID: \"8549fb56-bc75-4c66-8900-ba62a687ce0e\") " Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.176019 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-utilities" (OuterVolumeSpecName: "utilities") pod "8549fb56-bc75-4c66-8900-ba62a687ce0e" (UID: "8549fb56-bc75-4c66-8900-ba62a687ce0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.190005 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8549fb56-bc75-4c66-8900-ba62a687ce0e-kube-api-access-nrzgn" (OuterVolumeSpecName: "kube-api-access-nrzgn") pod "8549fb56-bc75-4c66-8900-ba62a687ce0e" (UID: "8549fb56-bc75-4c66-8900-ba62a687ce0e"). InnerVolumeSpecName "kube-api-access-nrzgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.233589 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8549fb56-bc75-4c66-8900-ba62a687ce0e" (UID: "8549fb56-bc75-4c66-8900-ba62a687ce0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.277919 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.277975 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrzgn\" (UniqueName: \"kubernetes.io/projected/8549fb56-bc75-4c66-8900-ba62a687ce0e-kube-api-access-nrzgn\") on node \"crc\" DevicePath \"\"" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.277996 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8549fb56-bc75-4c66-8900-ba62a687ce0e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.527601 4778 generic.go:334] "Generic (PLEG): container finished" podID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerID="fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3" exitCode=0 Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.527654 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerDied","Data":"fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3"} Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.527669 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdrpx" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.527698 4778 scope.go:117] "RemoveContainer" containerID="fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.527685 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdrpx" event={"ID":"8549fb56-bc75-4c66-8900-ba62a687ce0e","Type":"ContainerDied","Data":"74af807183905ac63573678f5f2eb3366ccf037a22980bf197b990c23186ec82"} Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.561073 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdrpx"] Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.569991 4778 scope.go:117] "RemoveContainer" containerID="808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.576204 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdrpx"] Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.599469 4778 scope.go:117] "RemoveContainer" containerID="9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.665879 4778 scope.go:117] "RemoveContainer" containerID="fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3" Mar 12 15:13:26 crc kubenswrapper[4778]: E0312 15:13:26.666571 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3\": container with ID starting with fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3 not found: ID does not exist" containerID="fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.666609 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3"} err="failed to get container status \"fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3\": rpc error: code = NotFound desc = could not find container \"fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3\": container with ID starting with fc619cbf16d89a1d02455527225255d2e2cdb8ec4ab739a1921cf7f4ffd16ae3 not found: ID does not exist" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.666634 4778 scope.go:117] "RemoveContainer" containerID="808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762" Mar 12 15:13:26 crc kubenswrapper[4778]: E0312 15:13:26.667100 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762\": container with ID starting with 808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762 not found: ID does not exist" containerID="808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.667132 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762"} err="failed to get container status \"808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762\": rpc error: code = NotFound desc = could not find container \"808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762\": container with ID starting with 808d468b4609825642a3f2c61ee5a962c7a86572772b7af2e05f7c2ee27fe762 not found: ID does not exist" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.667149 4778 scope.go:117] "RemoveContainer" containerID="9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8" Mar 12 15:13:26 crc kubenswrapper[4778]: E0312 15:13:26.667633 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8\": container with ID starting with 9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8 not found: ID does not exist" containerID="9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8" Mar 12 15:13:26 crc kubenswrapper[4778]: I0312 15:13:26.667660 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8"} err="failed to get container status \"9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8\": rpc error: code = NotFound desc = could not find container \"9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8\": container with ID starting with 9eba7d7cfb68560047065d50b50e472ab61f98cf3b77044f3ee6ed1c9751b8d8 not found: ID does not exist" Mar 12 15:13:28 crc kubenswrapper[4778]: I0312 15:13:28.266294 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" path="/var/lib/kubelet/pods/8549fb56-bc75-4c66-8900-ba62a687ce0e/volumes" Mar 12 15:13:35 crc kubenswrapper[4778]: I0312 15:13:35.254376 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:13:35 crc kubenswrapper[4778]: I0312 15:13:35.624119 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"9c0ffa691d48b1023164222bd8c69a88e4e7a89d268ba03833dc6ae4ab4b44b3"} Mar 12 15:13:47 crc kubenswrapper[4778]: I0312 15:13:47.675564 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/extract/0.log" Mar 12 15:13:58 crc kubenswrapper[4778]: I0312 15:13:58.975960 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-6h2c2_ffb8a1f4-4533-4368-a900-95d37fe1d3ad/manager/0.log" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.150001 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555474-s5qjz"] Mar 12 15:14:00 crc kubenswrapper[4778]: E0312 15:14:00.151120 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="extract-content" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151137 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="extract-content" Mar 12 15:14:00 crc kubenswrapper[4778]: E0312 15:14:00.151168 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="extract-utilities" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151214 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="extract-utilities" Mar 12 15:14:00 crc kubenswrapper[4778]: E0312 15:14:00.151234 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="registry-server" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151242 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="registry-server" Mar 12 15:14:00 crc kubenswrapper[4778]: E0312 15:14:00.151257 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="extract-content" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151289 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="extract-content" Mar 12 15:14:00 crc kubenswrapper[4778]: E0312 15:14:00.151319 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="extract-utilities" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151327 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="extract-utilities" Mar 12 15:14:00 crc kubenswrapper[4778]: E0312 15:14:00.151377 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="registry-server" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151387 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="registry-server" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151722 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="8549fb56-bc75-4c66-8900-ba62a687ce0e" containerName="registry-server" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.151785 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="690208cb-cbdb-488c-9998-70cf01f1cc05" containerName="registry-server" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.152938 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.155934 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.156330 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.156473 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.158159 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555474-s5qjz"] Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.272221 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2jq6\" (UniqueName: \"kubernetes.io/projected/c0be289e-351f-4101-acbd-0127a4b295dc-kube-api-access-v2jq6\") pod \"auto-csr-approver-29555474-s5qjz\" (UID: \"c0be289e-351f-4101-acbd-0127a4b295dc\") " pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.373751 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2jq6\" (UniqueName: \"kubernetes.io/projected/c0be289e-351f-4101-acbd-0127a4b295dc-kube-api-access-v2jq6\") pod \"auto-csr-approver-29555474-s5qjz\" (UID: \"c0be289e-351f-4101-acbd-0127a4b295dc\") " pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.397892 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2jq6\" (UniqueName: \"kubernetes.io/projected/c0be289e-351f-4101-acbd-0127a4b295dc-kube-api-access-v2jq6\") pod \"auto-csr-approver-29555474-s5qjz\" (UID: \"c0be289e-351f-4101-acbd-0127a4b295dc\") " pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.476833 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:00 crc kubenswrapper[4778]: I0312 15:14:00.957673 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555474-s5qjz"] Mar 12 15:14:01 crc kubenswrapper[4778]: I0312 15:14:01.282673 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-xm4cc_c8818ac0-af8b-42c9-a923-425fe79ed203/manager/0.log" Mar 12 15:14:01 crc kubenswrapper[4778]: I0312 15:14:01.715336 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-9n6jv_ad531191-d7c5-4ef6-9929-3a5869751d98/manager/0.log" Mar 12 15:14:01 crc kubenswrapper[4778]: I0312 15:14:01.889617 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" event={"ID":"c0be289e-351f-4101-acbd-0127a4b295dc","Type":"ContainerStarted","Data":"80ccd19c20c4529fd068dccf500b635d95b0d63bbb39a3c84b2d03bf76b7d944"} Mar 12 15:14:02 crc kubenswrapper[4778]: I0312 15:14:02.159290 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-gknp2_db7f6b97-2903-44bf-803f-c00c337400b9/manager/0.log" Mar 12 15:14:02 crc kubenswrapper[4778]: I0312 15:14:02.559223 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-b7tkm_e290c1ea-a39d-451e-a24b-17a2b61ff6f0/manager/0.log" Mar 12 15:14:02 crc kubenswrapper[4778]: I0312 15:14:02.901628 4778 generic.go:334] "Generic (PLEG): container finished" podID="c0be289e-351f-4101-acbd-0127a4b295dc" containerID="deb89f96ad2640fa0674d82f73344504fdcc846f9e4815ae8eef2ce9a216dca5" exitCode=0 Mar 12 15:14:02 crc kubenswrapper[4778]: I0312 15:14:02.901707 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" event={"ID":"c0be289e-351f-4101-acbd-0127a4b295dc","Type":"ContainerDied","Data":"deb89f96ad2640fa0674d82f73344504fdcc846f9e4815ae8eef2ce9a216dca5"} Mar 12 15:14:02 crc kubenswrapper[4778]: I0312 15:14:02.938941 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-4jgt8_4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71/manager/0.log" Mar 12 15:14:03 crc kubenswrapper[4778]: I0312 15:14:03.697480 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-5d6qz_02bc06ca-f4e6-4fde-bd5d-882714d9652c/manager/0.log" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.121717 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-qb8s8_98a4cfbd-3037-48b5-9047-5d574dcc0aca/manager/0.log" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.278208 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.377813 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2jq6\" (UniqueName: \"kubernetes.io/projected/c0be289e-351f-4101-acbd-0127a4b295dc-kube-api-access-v2jq6\") pod \"c0be289e-351f-4101-acbd-0127a4b295dc\" (UID: \"c0be289e-351f-4101-acbd-0127a4b295dc\") " Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.383456 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0be289e-351f-4101-acbd-0127a4b295dc-kube-api-access-v2jq6" (OuterVolumeSpecName: "kube-api-access-v2jq6") pod "c0be289e-351f-4101-acbd-0127a4b295dc" (UID: "c0be289e-351f-4101-acbd-0127a4b295dc"). InnerVolumeSpecName "kube-api-access-v2jq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.479591 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2jq6\" (UniqueName: \"kubernetes.io/projected/c0be289e-351f-4101-acbd-0127a4b295dc-kube-api-access-v2jq6\") on node \"crc\" DevicePath \"\"" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.614600 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7dxdh_7e02c37f-b9af-46c9-a743-03ead9b060db/manager/0.log" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.925961 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" event={"ID":"c0be289e-351f-4101-acbd-0127a4b295dc","Type":"ContainerDied","Data":"80ccd19c20c4529fd068dccf500b635d95b0d63bbb39a3c84b2d03bf76b7d944"} Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.926300 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80ccd19c20c4529fd068dccf500b635d95b0d63bbb39a3c84b2d03bf76b7d944" Mar 12 15:14:04 crc kubenswrapper[4778]: I0312 15:14:04.926040 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555474-s5qjz" Mar 12 15:14:05 crc kubenswrapper[4778]: I0312 15:14:05.031685 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-pn8tk_5e38a4fd-95f8-437b-923b-eca33b1387e6/manager/0.log" Mar 12 15:14:05 crc kubenswrapper[4778]: I0312 15:14:05.356545 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555468-tpk68"] Mar 12 15:14:05 crc kubenswrapper[4778]: I0312 15:14:05.366108 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555468-tpk68"] Mar 12 15:14:05 crc kubenswrapper[4778]: I0312 15:14:05.445306 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-jlbft_2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f/manager/0.log" Mar 12 15:14:05 crc kubenswrapper[4778]: I0312 15:14:05.887978 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-dd2ft_076835c9-352b-4e40-80c4-3bce3bb80594/manager/0.log" Mar 12 15:14:06 crc kubenswrapper[4778]: I0312 15:14:06.271667 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1760b72d-ab0b-489f-b263-7279ce51dc5f" path="/var/lib/kubelet/pods/1760b72d-ab0b-489f-b263-7279ce51dc5f/volumes" Mar 12 15:14:06 crc kubenswrapper[4778]: I0312 15:14:06.423472 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-686d5f9fbd-vv9rc_d7288cc6-4247-4d03-bd37-9862243bf613/manager/0.log" Mar 12 15:14:06 crc kubenswrapper[4778]: I0312 15:14:06.879429 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-cdgg9_1a01d06c-be6f-45de-a22d-c8f1058a3a84/manager/0.log" Mar 12 15:14:07 crc kubenswrapper[4778]: I0312 15:14:07.287371 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6_4f7d316e-6896-4f84-8423-6f79778c1c6b/manager/0.log" Mar 12 15:14:07 crc kubenswrapper[4778]: I0312 15:14:07.935276 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-5bc4df7446-x9bsl_34bbdc16-4518-4ee5-9a70-3cedcc5f0159/operator/0.log" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.100486 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ptp4v"] Mar 12 15:14:09 crc kubenswrapper[4778]: E0312 15:14:09.101449 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0be289e-351f-4101-acbd-0127a4b295dc" containerName="oc" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.101473 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0be289e-351f-4101-acbd-0127a4b295dc" containerName="oc" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.103787 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0be289e-351f-4101-acbd-0127a4b295dc" containerName="oc" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.108486 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.119713 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ptp4v"] Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.269412 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m98t\" (UniqueName: \"kubernetes.io/projected/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-kube-api-access-6m98t\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.269579 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-utilities\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.269697 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-catalog-content\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.370979 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-utilities\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.371107 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-catalog-content\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.371303 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m98t\" (UniqueName: \"kubernetes.io/projected/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-kube-api-access-6m98t\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.371616 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-catalog-content\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.372253 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-utilities\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.390537 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m98t\" (UniqueName: \"kubernetes.io/projected/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-kube-api-access-6m98t\") pod \"redhat-operators-ptp4v\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.434856 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.588024 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5785b7957-7vdgw_d0784623-5f08-4109-9c7e-0a329210ce07/manager/0.log" Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.931775 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ptp4v"] Mar 12 15:14:09 crc kubenswrapper[4778]: I0312 15:14:09.970571 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerStarted","Data":"04acc9cfa9e3d6a97a649b3f446aec567265c61efd313e0db14406ccbd638aff"} Mar 12 15:14:10 crc kubenswrapper[4778]: I0312 15:14:10.044309 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-b2fsv_748546a6-1355-470f-b8d0-de395cf3f681/registry-server/0.log" Mar 12 15:14:10 crc kubenswrapper[4778]: I0312 15:14:10.456243 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-bbgmb_8d38fd7e-6fa1-4b0c-9c82-9c57290c7837/manager/0.log" Mar 12 15:14:10 crc kubenswrapper[4778]: I0312 15:14:10.902235 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-wvpf8_52524252-25bd-49e5-822e-3d4668aff2f9/manager/0.log" Mar 12 15:14:10 crc kubenswrapper[4778]: I0312 15:14:10.979795 4778 generic.go:334] "Generic (PLEG): container finished" podID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerID="217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e" exitCode=0 Mar 12 15:14:10 crc kubenswrapper[4778]: I0312 15:14:10.979832 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerDied","Data":"217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e"} Mar 12 15:14:11 crc kubenswrapper[4778]: I0312 15:14:11.316613 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-shf7b_034f39d8-a33e-4e37-bcde-51fb22debdd1/operator/0.log" Mar 12 15:14:11 crc kubenswrapper[4778]: I0312 15:14:11.754843 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-84mps_64a36384-f2e6-4077-b2ca-de2a6ce6ea06/manager/0.log" Mar 12 15:14:11 crc kubenswrapper[4778]: I0312 15:14:11.993098 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerStarted","Data":"dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1"} Mar 12 15:14:12 crc kubenswrapper[4778]: I0312 15:14:12.237150 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-gfv5z_6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c/manager/0.log" Mar 12 15:14:12 crc kubenswrapper[4778]: I0312 15:14:12.687440 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-pcfrz_ed9b9271-4ae9-440a-9411-15d46267106e/manager/0.log" Mar 12 15:14:13 crc kubenswrapper[4778]: I0312 15:14:13.143687 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-2tjsk_8c02ecb8-0e15-4672-823a-c4437ca5bf8c/manager/0.log" Mar 12 15:14:15 crc kubenswrapper[4778]: I0312 15:14:15.030625 4778 generic.go:334] "Generic (PLEG): container finished" podID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerID="dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1" exitCode=0 Mar 12 15:14:15 crc kubenswrapper[4778]: I0312 15:14:15.030705 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerDied","Data":"dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1"} Mar 12 15:14:16 crc kubenswrapper[4778]: I0312 15:14:16.040949 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerStarted","Data":"d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2"} Mar 12 15:14:16 crc kubenswrapper[4778]: I0312 15:14:16.060820 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ptp4v" podStartSLOduration=2.479271506 podStartE2EDuration="7.060801443s" podCreationTimestamp="2026-03-12 15:14:09 +0000 UTC" firstStartedPulling="2026-03-12 15:14:10.982500545 +0000 UTC m=+7469.431195941" lastFinishedPulling="2026-03-12 15:14:15.564030472 +0000 UTC m=+7474.012725878" observedRunningTime="2026-03-12 15:14:16.060632608 +0000 UTC m=+7474.509328004" watchObservedRunningTime="2026-03-12 15:14:16.060801443 +0000 UTC m=+7474.509496839" Mar 12 15:14:18 crc kubenswrapper[4778]: I0312 15:14:18.081756 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-86cb765474-5pq5z_6bd172c5-383f-4273-98a5-2c92223dc765/barbican-api-log/0.log" Mar 12 15:14:18 crc kubenswrapper[4778]: I0312 15:14:18.910611 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65c9994dfd-xznqh_8ee1f546-8428-4b23-93e4-b8370fd4224b/barbican-keystone-listener-log/0.log" Mar 12 15:14:19 crc kubenswrapper[4778]: I0312 15:14:19.435482 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:19 crc kubenswrapper[4778]: I0312 15:14:19.435535 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:19 crc kubenswrapper[4778]: I0312 15:14:19.463797 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7dcf9787-ngc87_d505bb59-3c9e-4cfa-891c-c8e0068e2567/barbican-worker-log/0.log" Mar 12 15:14:20 crc kubenswrapper[4778]: I0312 15:14:20.040216 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx_b99627a8-43d8-4f7d-90f7-530eda3c2213/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:20 crc kubenswrapper[4778]: I0312 15:14:20.501223 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ptp4v" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="registry-server" probeResult="failure" output=< Mar 12 15:14:20 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 15:14:20 crc kubenswrapper[4778]: > Mar 12 15:14:20 crc kubenswrapper[4778]: I0312 15:14:20.663057 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/ceilometer-central-agent/0.log" Mar 12 15:14:21 crc kubenswrapper[4778]: I0312 15:14:21.160743 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_99f72014-50e8-4dd4-9764-1b2c7d546b30/cinder-api-log/0.log" Mar 12 15:14:21 crc kubenswrapper[4778]: I0312 15:14:21.693820 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_39ee2404-53a8-4598-8c4b-c3a34fbf3480/cinder-scheduler/0.log" Mar 12 15:14:22 crc kubenswrapper[4778]: I0312 15:14:22.222812 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4szjl_5c5541f3-fb44-476b-91c2-b07dffe50894/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:22 crc kubenswrapper[4778]: I0312 15:14:22.750021 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6_36bb4acd-fab3-4998-a8cd-a6ebcc800fc8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:24 crc kubenswrapper[4778]: I0312 15:14:24.221130 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/dnsmasq-dns/0.log" Mar 12 15:14:24 crc kubenswrapper[4778]: I0312 15:14:24.731314 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-2xksx_96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:25 crc kubenswrapper[4778]: I0312 15:14:25.171090 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_81c1a05c-5642-43d4-8a7b-229330168332/glance-log/0.log" Mar 12 15:14:25 crc kubenswrapper[4778]: I0312 15:14:25.654587 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7fa757af-1c91-4b93-8916-5bbd99b8522e/glance-log/0.log" Mar 12 15:14:26 crc kubenswrapper[4778]: I0312 15:14:26.119839 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bngcx_f69e6cfe-f7c2-4127-b4df-710725c52227/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:26 crc kubenswrapper[4778]: I0312 15:14:26.504345 4778 scope.go:117] "RemoveContainer" containerID="98229a921540253e11a1f90e715794eabbd7a1c19952afa6969c95cd62f6a069" Mar 12 15:14:26 crc kubenswrapper[4778]: I0312 15:14:26.755838 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g252n_29f8609b-4a3b-42ba-9450-a2b633bb4c2c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:28 crc kubenswrapper[4778]: I0312 15:14:28.358608 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-6lrlq_a56bb599-f10d-4564-b6bf-48128dc2c7f1/keystone-api/0.log" Mar 12 15:14:29 crc kubenswrapper[4778]: I0312 15:14:29.486619 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:29 crc kubenswrapper[4778]: I0312 15:14:29.542158 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:29 crc kubenswrapper[4778]: I0312 15:14:29.720636 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ptp4v"] Mar 12 15:14:29 crc kubenswrapper[4778]: I0312 15:14:29.887281 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-z4h9m_16dea17b-eaa4-4bbf-8895-c077b3e28d66/keystone-api/0.log" Mar 12 15:14:30 crc kubenswrapper[4778]: I0312 15:14:30.353541 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555401-vjgkl_e4df6927-3452-4b36-b59a-a1fdcd4272a4/keystone-cron/0.log" Mar 12 15:14:30 crc kubenswrapper[4778]: I0312 15:14:30.821534 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555461-lmqk9_ebdf3274-70cb-4083-bf12-5d1038a9b7ba/keystone-cron/0.log" Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.188156 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ptp4v" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="registry-server" containerID="cri-o://d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2" gracePeriod=2 Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.266031 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_51f24fcd-aff5-4785-abf7-4936180cee78/kube-state-metrics/0.log" Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.729835 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8_8713b951-b516-42bd-9286-4343e5bcc955/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.739654 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.917711 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-catalog-content\") pod \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.917904 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-utilities\") pod \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.917966 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m98t\" (UniqueName: \"kubernetes.io/projected/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-kube-api-access-6m98t\") pod \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\" (UID: \"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d\") " Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.918971 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-utilities" (OuterVolumeSpecName: "utilities") pod "0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" (UID: "0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:14:31 crc kubenswrapper[4778]: I0312 15:14:31.924389 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-kube-api-access-6m98t" (OuterVolumeSpecName: "kube-api-access-6m98t") pod "0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" (UID: "0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d"). InnerVolumeSpecName "kube-api-access-6m98t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.020731 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.020777 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m98t\" (UniqueName: \"kubernetes.io/projected/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-kube-api-access-6m98t\") on node \"crc\" DevicePath \"\"" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.083993 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" (UID: "0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.122198 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.205683 4778 generic.go:334] "Generic (PLEG): container finished" podID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerID="d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2" exitCode=0 Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.205737 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerDied","Data":"d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2"} Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.205753 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptp4v" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.205775 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptp4v" event={"ID":"0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d","Type":"ContainerDied","Data":"04acc9cfa9e3d6a97a649b3f446aec567265c61efd313e0db14406ccbd638aff"} Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.205802 4778 scope.go:117] "RemoveContainer" containerID="d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.257482 4778 scope.go:117] "RemoveContainer" containerID="dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.280744 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ptp4v"] Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.286435 4778 scope.go:117] "RemoveContainer" containerID="217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.287664 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ptp4v"] Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.337465 4778 scope.go:117] "RemoveContainer" containerID="d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2" Mar 12 15:14:32 crc kubenswrapper[4778]: E0312 15:14:32.338044 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2\": container with ID starting with d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2 not found: ID does not exist" containerID="d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.338101 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2"} err="failed to get container status \"d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2\": rpc error: code = NotFound desc = could not find container \"d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2\": container with ID starting with d16e27ae8470c7352f41095add77058c0bdaf9645d86b1b180d532da10b9aee2 not found: ID does not exist" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.338138 4778 scope.go:117] "RemoveContainer" containerID="dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1" Mar 12 15:14:32 crc kubenswrapper[4778]: E0312 15:14:32.338585 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1\": container with ID starting with dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1 not found: ID does not exist" containerID="dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.338649 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1"} err="failed to get container status \"dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1\": rpc error: code = NotFound desc = could not find container \"dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1\": container with ID starting with dc5279aa56d851d1367ade7841dcd4e8d413b3295acaa8381ae6f7bf32d030c1 not found: ID does not exist" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.338679 4778 scope.go:117] "RemoveContainer" containerID="217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e" Mar 12 15:14:32 crc kubenswrapper[4778]: E0312 15:14:32.339127 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e\": container with ID starting with 217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e not found: ID does not exist" containerID="217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e" Mar 12 15:14:32 crc kubenswrapper[4778]: I0312 15:14:32.339216 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e"} err="failed to get container status \"217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e\": rpc error: code = NotFound desc = could not find container \"217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e\": container with ID starting with 217a23ca1a727d114e2cf1fc5f98b8b425bc25bc2bf661e3774449d8591c7a7e not found: ID does not exist" Mar 12 15:14:33 crc kubenswrapper[4778]: I0312 15:14:33.115439 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ec63cc68-6fde-419b-973c-91fc982e6a49/memcached/0.log" Mar 12 15:14:34 crc kubenswrapper[4778]: I0312 15:14:34.272143 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" path="/var/lib/kubelet/pods/0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d/volumes" Mar 12 15:14:37 crc kubenswrapper[4778]: I0312 15:14:37.321749 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-dggmh_7596a69e-33c9-4a2b-89fc-e4c41252b3fd/neutron-api/0.log" Mar 12 15:14:41 crc kubenswrapper[4778]: I0312 15:14:41.483256 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-zx97x_8a67d4b7-d8eb-40f4-b51d-62e92c6042c1/neutron-api/0.log" Mar 12 15:14:41 crc kubenswrapper[4778]: I0312 15:14:41.987659 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg_5cc410de-5b42-44d1-8b29-37161475730e/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:44 crc kubenswrapper[4778]: I0312 15:14:44.021790 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4/nova-api-log/0.log" Mar 12 15:14:45 crc kubenswrapper[4778]: I0312 15:14:45.753926 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-1_f0341d80-4327-4c9e-bc11-0cddbc6eab66/nova-api-log/0.log" Mar 12 15:14:46 crc kubenswrapper[4778]: I0312 15:14:46.792287 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_929bb450-949d-4f4f-9c21-de6c3fe32927/nova-cell0-conductor-conductor/0.log" Mar 12 15:14:47 crc kubenswrapper[4778]: I0312 15:14:47.516844 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1466aea3-fa10-49a6-a254-a96a52091aca/nova-cell1-conductor-conductor/0.log" Mar 12 15:14:48 crc kubenswrapper[4778]: I0312 15:14:48.098103 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-metadata-0_c289a520-78eb-433f-b7a4-0c03be917c18/nova-cell1-metadata-log/0.log" Mar 12 15:14:48 crc kubenswrapper[4778]: I0312 15:14:48.686627 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7/nova-cell1-novncproxy-novncproxy/0.log" Mar 12 15:14:49 crc kubenswrapper[4778]: I0312 15:14:49.285527 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5tw6s_6ed77f87-e6b2-4c7a-8b0e-003106200dc8/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:50 crc kubenswrapper[4778]: I0312 15:14:50.686525 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f613745b-fe33-4918-9e0a-da2a59c55e33/nova-scheduler-scheduler/0.log" Mar 12 15:14:51 crc kubenswrapper[4778]: I0312 15:14:51.157221 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/galera/0.log" Mar 12 15:14:51 crc kubenswrapper[4778]: I0312 15:14:51.639156 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/galera/0.log" Mar 12 15:14:52 crc kubenswrapper[4778]: I0312 15:14:52.076311 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_856cd6d1-db21-4503-94d7-cbf27ca96cc2/openstackclient/0.log" Mar 12 15:14:52 crc kubenswrapper[4778]: I0312 15:14:52.576565 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4wct6_3b8efd1e-884d-4963-b69f-04ede0a92267/ovn-controller/0.log" Mar 12 15:14:53 crc kubenswrapper[4778]: I0312 15:14:53.052343 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vtt4z_a8484e5d-6f77-407c-81db-0d9b2a6b37fd/openstack-network-exporter/0.log" Mar 12 15:14:53 crc kubenswrapper[4778]: I0312 15:14:53.516625 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server/0.log" Mar 12 15:14:53 crc kubenswrapper[4778]: I0312 15:14:53.990447 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9lbdq_3c0a2200-506d-4ac3-b08c-9b3156c9e573/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:54 crc kubenswrapper[4778]: I0312 15:14:54.417021 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b25f9c9-784a-4a52-9bb3-02c6c4592702/ovn-northd/0.log" Mar 12 15:14:54 crc kubenswrapper[4778]: I0312 15:14:54.833176 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7321e15e-673c-4e0d-80f8-6ac644c1940f/ovsdbserver-nb/0.log" Mar 12 15:14:55 crc kubenswrapper[4778]: I0312 15:14:55.293024 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c951c6f-06fd-4793-a95b-26b5c1400d73/ovsdbserver-sb/0.log" Mar 12 15:14:56 crc kubenswrapper[4778]: I0312 15:14:56.245385 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d4d765698-l7bjx_267e7df2-d35c-45c4-af65-e8af31f8f6cf/placement-log/0.log" Mar 12 15:14:56 crc kubenswrapper[4778]: I0312 15:14:56.860976 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/rabbitmq/0.log" Mar 12 15:14:57 crc kubenswrapper[4778]: I0312 15:14:57.416625 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/rabbitmq/0.log" Mar 12 15:14:57 crc kubenswrapper[4778]: I0312 15:14:57.922709 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc_43a3ffe4-8b64-4e26-b63a-5254a986e4a4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:58 crc kubenswrapper[4778]: I0312 15:14:58.366997 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc_bd7ac6b4-5600-45ce-b0ea-199dd4baefcb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:58 crc kubenswrapper[4778]: I0312 15:14:58.803691 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gt58t_b0bb06df-44bb-4939-9492-a6ad3d6b5368/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:14:59 crc kubenswrapper[4778]: I0312 15:14:59.238754 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8mmjm_c993b33e-6c36-4524-864a-65da461a8e0c/ssh-known-hosts-edpm-deployment/0.log" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.006947 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-77f887c49f-fw2qd_bbd76cb8-462f-4e60-b755-ef3170e70d11/proxy-httpd/0.log" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.147525 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k"] Mar 12 15:15:00 crc kubenswrapper[4778]: E0312 15:15:00.148176 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="extract-utilities" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.148271 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="extract-utilities" Mar 12 15:15:00 crc kubenswrapper[4778]: E0312 15:15:00.148350 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="registry-server" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.148407 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="registry-server" Mar 12 15:15:00 crc kubenswrapper[4778]: E0312 15:15:00.148505 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="extract-content" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.148821 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="extract-content" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.149066 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6c3f2d-9c9f-406c-86ba-1eee5ef9228d" containerName="registry-server" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.149819 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.152602 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.152866 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.159396 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k"] Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.292536 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f92778df-5bc8-42ec-b5c7-6f938cefef60-config-volume\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.292629 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f92778df-5bc8-42ec-b5c7-6f938cefef60-secret-volume\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.292785 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsgwl\" (UniqueName: \"kubernetes.io/projected/f92778df-5bc8-42ec-b5c7-6f938cefef60-kube-api-access-jsgwl\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.394135 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsgwl\" (UniqueName: \"kubernetes.io/projected/f92778df-5bc8-42ec-b5c7-6f938cefef60-kube-api-access-jsgwl\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.394223 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f92778df-5bc8-42ec-b5c7-6f938cefef60-config-volume\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.394302 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f92778df-5bc8-42ec-b5c7-6f938cefef60-secret-volume\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.398124 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f92778df-5bc8-42ec-b5c7-6f938cefef60-config-volume\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.402024 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f92778df-5bc8-42ec-b5c7-6f938cefef60-secret-volume\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.416764 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsgwl\" (UniqueName: \"kubernetes.io/projected/f92778df-5bc8-42ec-b5c7-6f938cefef60-kube-api-access-jsgwl\") pod \"collect-profiles-29555475-ll84k\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.455605 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5knbg_2edc2c90-f91e-402d-809c-514e9d8a5e04/swift-ring-rebalance/0.log" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.480009 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.866202 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-server/0.log" Mar 12 15:15:00 crc kubenswrapper[4778]: I0312 15:15:00.962379 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k"] Mar 12 15:15:01 crc kubenswrapper[4778]: I0312 15:15:01.342118 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s_2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:15:01 crc kubenswrapper[4778]: I0312 15:15:01.533540 4778 generic.go:334] "Generic (PLEG): container finished" podID="f92778df-5bc8-42ec-b5c7-6f938cefef60" containerID="851bf80a56ed9718274664a37b6b3f8f6a2cbf8dbba58e45165c53b43e774224" exitCode=0 Mar 12 15:15:01 crc kubenswrapper[4778]: I0312 15:15:01.533587 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" event={"ID":"f92778df-5bc8-42ec-b5c7-6f938cefef60","Type":"ContainerDied","Data":"851bf80a56ed9718274664a37b6b3f8f6a2cbf8dbba58e45165c53b43e774224"} Mar 12 15:15:01 crc kubenswrapper[4778]: I0312 15:15:01.533612 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" event={"ID":"f92778df-5bc8-42ec-b5c7-6f938cefef60","Type":"ContainerStarted","Data":"2384e90392d7e66ed75d93be62b562a34f6b3f71f59ea646143b70c89b11cc75"} Mar 12 15:15:01 crc kubenswrapper[4778]: I0312 15:15:01.804970 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_74897d0a-ca7b-4589-bd4c-75910c2d491c/tempest-tests-tempest-tests-runner/0.log" Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.261476 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_82246f69-2112-44e9-a783-a4a5926188b4/test-operator-logs-container/0.log" Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.723374 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9glvr_41583476-38cd-4c0d-a05a-96ddc5b330ca/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.825317 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.945550 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f92778df-5bc8-42ec-b5c7-6f938cefef60-secret-volume\") pod \"f92778df-5bc8-42ec-b5c7-6f938cefef60\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.945610 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsgwl\" (UniqueName: \"kubernetes.io/projected/f92778df-5bc8-42ec-b5c7-6f938cefef60-kube-api-access-jsgwl\") pod \"f92778df-5bc8-42ec-b5c7-6f938cefef60\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.945750 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f92778df-5bc8-42ec-b5c7-6f938cefef60-config-volume\") pod \"f92778df-5bc8-42ec-b5c7-6f938cefef60\" (UID: \"f92778df-5bc8-42ec-b5c7-6f938cefef60\") " Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.947058 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92778df-5bc8-42ec-b5c7-6f938cefef60-config-volume" (OuterVolumeSpecName: "config-volume") pod "f92778df-5bc8-42ec-b5c7-6f938cefef60" (UID: "f92778df-5bc8-42ec-b5c7-6f938cefef60"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.957591 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92778df-5bc8-42ec-b5c7-6f938cefef60-kube-api-access-jsgwl" (OuterVolumeSpecName: "kube-api-access-jsgwl") pod "f92778df-5bc8-42ec-b5c7-6f938cefef60" (UID: "f92778df-5bc8-42ec-b5c7-6f938cefef60"). InnerVolumeSpecName "kube-api-access-jsgwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:15:02 crc kubenswrapper[4778]: I0312 15:15:02.964161 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92778df-5bc8-42ec-b5c7-6f938cefef60-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f92778df-5bc8-42ec-b5c7-6f938cefef60" (UID: "f92778df-5bc8-42ec-b5c7-6f938cefef60"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.049162 4778 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f92778df-5bc8-42ec-b5c7-6f938cefef60-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.049311 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsgwl\" (UniqueName: \"kubernetes.io/projected/f92778df-5bc8-42ec-b5c7-6f938cefef60-kube-api-access-jsgwl\") on node \"crc\" DevicePath \"\"" Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.049326 4778 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f92778df-5bc8-42ec-b5c7-6f938cefef60-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.555236 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" event={"ID":"f92778df-5bc8-42ec-b5c7-6f938cefef60","Type":"ContainerDied","Data":"2384e90392d7e66ed75d93be62b562a34f6b3f71f59ea646143b70c89b11cc75"} Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.555302 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2384e90392d7e66ed75d93be62b562a34f6b3f71f59ea646143b70c89b11cc75" Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.555390 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555475-ll84k" Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.906308 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl"] Mar 12 15:15:03 crc kubenswrapper[4778]: I0312 15:15:03.906589 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555430-zhqfl"] Mar 12 15:15:04 crc kubenswrapper[4778]: I0312 15:15:04.264273 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4d57b8-99e5-4955-a6fe-9b0c0a6e61df" path="/var/lib/kubelet/pods/db4d57b8-99e5-4955-a6fe-9b0c0a6e61df/volumes" Mar 12 15:15:26 crc kubenswrapper[4778]: I0312 15:15:26.625491 4778 scope.go:117] "RemoveContainer" containerID="57a58448ac2691d1255487422cd2ce72ba1abcb298bf6c4ed12464fdb32a532d" Mar 12 15:15:40 crc kubenswrapper[4778]: I0312 15:15:40.028890 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/extract/0.log" Mar 12 15:15:51 crc kubenswrapper[4778]: I0312 15:15:51.116980 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-6h2c2_ffb8a1f4-4533-4368-a900-95d37fe1d3ad/manager/0.log" Mar 12 15:15:53 crc kubenswrapper[4778]: I0312 15:15:53.400351 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-xm4cc_c8818ac0-af8b-42c9-a923-425fe79ed203/manager/0.log" Mar 12 15:15:53 crc kubenswrapper[4778]: I0312 15:15:53.807584 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-9n6jv_ad531191-d7c5-4ef6-9929-3a5869751d98/manager/0.log" Mar 12 15:15:54 crc kubenswrapper[4778]: I0312 15:15:54.254010 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-gknp2_db7f6b97-2903-44bf-803f-c00c337400b9/manager/0.log" Mar 12 15:15:54 crc kubenswrapper[4778]: I0312 15:15:54.758555 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-b7tkm_e290c1ea-a39d-451e-a24b-17a2b61ff6f0/manager/0.log" Mar 12 15:15:55 crc kubenswrapper[4778]: I0312 15:15:55.143966 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-4jgt8_4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71/manager/0.log" Mar 12 15:15:55 crc kubenswrapper[4778]: I0312 15:15:55.796459 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-5d6qz_02bc06ca-f4e6-4fde-bd5d-882714d9652c/manager/0.log" Mar 12 15:15:56 crc kubenswrapper[4778]: I0312 15:15:56.229654 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-qb8s8_98a4cfbd-3037-48b5-9047-5d574dcc0aca/manager/0.log" Mar 12 15:15:56 crc kubenswrapper[4778]: I0312 15:15:56.676502 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7dxdh_7e02c37f-b9af-46c9-a743-03ead9b060db/manager/0.log" Mar 12 15:15:57 crc kubenswrapper[4778]: I0312 15:15:57.132574 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-pn8tk_5e38a4fd-95f8-437b-923b-eca33b1387e6/manager/0.log" Mar 12 15:15:57 crc kubenswrapper[4778]: I0312 15:15:57.522624 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-jlbft_2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f/manager/0.log" Mar 12 15:15:57 crc kubenswrapper[4778]: I0312 15:15:57.990046 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-dd2ft_076835c9-352b-4e40-80c4-3bce3bb80594/manager/0.log" Mar 12 15:15:58 crc kubenswrapper[4778]: I0312 15:15:58.484964 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-686d5f9fbd-vv9rc_d7288cc6-4247-4d03-bd37-9862243bf613/manager/0.log" Mar 12 15:15:58 crc kubenswrapper[4778]: I0312 15:15:58.557663 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:15:58 crc kubenswrapper[4778]: I0312 15:15:58.557725 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:15:58 crc kubenswrapper[4778]: I0312 15:15:58.879094 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-cdgg9_1a01d06c-be6f-45de-a22d-c8f1058a3a84/manager/0.log" Mar 12 15:15:59 crc kubenswrapper[4778]: I0312 15:15:59.249577 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6_4f7d316e-6896-4f84-8423-6f79778c1c6b/manager/0.log" Mar 12 15:15:59 crc kubenswrapper[4778]: I0312 15:15:59.734359 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-5bc4df7446-x9bsl_34bbdc16-4518-4ee5-9a70-3cedcc5f0159/operator/0.log" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.158799 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555476-5dkdj"] Mar 12 15:16:00 crc kubenswrapper[4778]: E0312 15:16:00.159866 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92778df-5bc8-42ec-b5c7-6f938cefef60" containerName="collect-profiles" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.159961 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92778df-5bc8-42ec-b5c7-6f938cefef60" containerName="collect-profiles" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.160291 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92778df-5bc8-42ec-b5c7-6f938cefef60" containerName="collect-profiles" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.160996 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.163532 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.163842 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.163884 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.180058 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555476-5dkdj"] Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.205453 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r2zc\" (UniqueName: \"kubernetes.io/projected/e42af641-c33e-4b80-899f-98e5d4e78dad-kube-api-access-4r2zc\") pod \"auto-csr-approver-29555476-5dkdj\" (UID: \"e42af641-c33e-4b80-899f-98e5d4e78dad\") " pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.307421 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r2zc\" (UniqueName: \"kubernetes.io/projected/e42af641-c33e-4b80-899f-98e5d4e78dad-kube-api-access-4r2zc\") pod \"auto-csr-approver-29555476-5dkdj\" (UID: \"e42af641-c33e-4b80-899f-98e5d4e78dad\") " pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.328611 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r2zc\" (UniqueName: \"kubernetes.io/projected/e42af641-c33e-4b80-899f-98e5d4e78dad-kube-api-access-4r2zc\") pod \"auto-csr-approver-29555476-5dkdj\" (UID: \"e42af641-c33e-4b80-899f-98e5d4e78dad\") " pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.484086 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.952235 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555476-5dkdj"] Mar 12 15:16:00 crc kubenswrapper[4778]: I0312 15:16:00.966343 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 15:16:01 crc kubenswrapper[4778]: I0312 15:16:01.139890 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" event={"ID":"e42af641-c33e-4b80-899f-98e5d4e78dad","Type":"ContainerStarted","Data":"993bb09d3a882b5a327b311cfd7cd720701e36ad48212d0f63e2c927ffc5b5ff"} Mar 12 15:16:01 crc kubenswrapper[4778]: I0312 15:16:01.465565 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5785b7957-7vdgw_d0784623-5f08-4109-9c7e-0a329210ce07/manager/0.log" Mar 12 15:16:01 crc kubenswrapper[4778]: I0312 15:16:01.863901 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-b2fsv_748546a6-1355-470f-b8d0-de395cf3f681/registry-server/0.log" Mar 12 15:16:02 crc kubenswrapper[4778]: I0312 15:16:02.288899 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-bbgmb_8d38fd7e-6fa1-4b0c-9c82-9c57290c7837/manager/0.log" Mar 12 15:16:02 crc kubenswrapper[4778]: I0312 15:16:02.722176 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-wvpf8_52524252-25bd-49e5-822e-3d4668aff2f9/manager/0.log" Mar 12 15:16:03 crc kubenswrapper[4778]: I0312 15:16:03.146256 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-shf7b_034f39d8-a33e-4e37-bcde-51fb22debdd1/operator/0.log" Mar 12 15:16:03 crc kubenswrapper[4778]: I0312 15:16:03.169960 4778 generic.go:334] "Generic (PLEG): container finished" podID="e42af641-c33e-4b80-899f-98e5d4e78dad" containerID="9e696920a26a473d829a12f6bf276893531dc1bd498cf28bdf23c8b663c144ee" exitCode=0 Mar 12 15:16:03 crc kubenswrapper[4778]: I0312 15:16:03.170008 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" event={"ID":"e42af641-c33e-4b80-899f-98e5d4e78dad","Type":"ContainerDied","Data":"9e696920a26a473d829a12f6bf276893531dc1bd498cf28bdf23c8b663c144ee"} Mar 12 15:16:03 crc kubenswrapper[4778]: I0312 15:16:03.520998 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-84mps_64a36384-f2e6-4077-b2ca-de2a6ce6ea06/manager/0.log" Mar 12 15:16:03 crc kubenswrapper[4778]: I0312 15:16:03.937091 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-gfv5z_6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c/manager/0.log" Mar 12 15:16:04 crc kubenswrapper[4778]: I0312 15:16:04.363165 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-pcfrz_ed9b9271-4ae9-440a-9411-15d46267106e/manager/0.log" Mar 12 15:16:04 crc kubenswrapper[4778]: I0312 15:16:04.516016 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:04 crc kubenswrapper[4778]: I0312 15:16:04.594791 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r2zc\" (UniqueName: \"kubernetes.io/projected/e42af641-c33e-4b80-899f-98e5d4e78dad-kube-api-access-4r2zc\") pod \"e42af641-c33e-4b80-899f-98e5d4e78dad\" (UID: \"e42af641-c33e-4b80-899f-98e5d4e78dad\") " Mar 12 15:16:04 crc kubenswrapper[4778]: I0312 15:16:04.601522 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e42af641-c33e-4b80-899f-98e5d4e78dad-kube-api-access-4r2zc" (OuterVolumeSpecName: "kube-api-access-4r2zc") pod "e42af641-c33e-4b80-899f-98e5d4e78dad" (UID: "e42af641-c33e-4b80-899f-98e5d4e78dad"). InnerVolumeSpecName "kube-api-access-4r2zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:16:04 crc kubenswrapper[4778]: I0312 15:16:04.697428 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r2zc\" (UniqueName: \"kubernetes.io/projected/e42af641-c33e-4b80-899f-98e5d4e78dad-kube-api-access-4r2zc\") on node \"crc\" DevicePath \"\"" Mar 12 15:16:04 crc kubenswrapper[4778]: I0312 15:16:04.736597 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-2tjsk_8c02ecb8-0e15-4672-823a-c4437ca5bf8c/manager/0.log" Mar 12 15:16:05 crc kubenswrapper[4778]: I0312 15:16:05.189555 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" event={"ID":"e42af641-c33e-4b80-899f-98e5d4e78dad","Type":"ContainerDied","Data":"993bb09d3a882b5a327b311cfd7cd720701e36ad48212d0f63e2c927ffc5b5ff"} Mar 12 15:16:05 crc kubenswrapper[4778]: I0312 15:16:05.189935 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="993bb09d3a882b5a327b311cfd7cd720701e36ad48212d0f63e2c927ffc5b5ff" Mar 12 15:16:05 crc kubenswrapper[4778]: I0312 15:16:05.189748 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555476-5dkdj" Mar 12 15:16:05 crc kubenswrapper[4778]: I0312 15:16:05.592948 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555470-65l68"] Mar 12 15:16:05 crc kubenswrapper[4778]: I0312 15:16:05.603677 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555470-65l68"] Mar 12 15:16:06 crc kubenswrapper[4778]: I0312 15:16:06.265092 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f05c65b6-74a0-49ef-8f84-3b4453313dc7" path="/var/lib/kubelet/pods/f05c65b6-74a0-49ef-8f84-3b4453313dc7/volumes" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.921959 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t8fhv/must-gather-wpn7c"] Mar 12 15:16:25 crc kubenswrapper[4778]: E0312 15:16:25.926822 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e42af641-c33e-4b80-899f-98e5d4e78dad" containerName="oc" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.926847 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="e42af641-c33e-4b80-899f-98e5d4e78dad" containerName="oc" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.927119 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="e42af641-c33e-4b80-899f-98e5d4e78dad" containerName="oc" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.928484 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.938047 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t8fhv"/"openshift-service-ca.crt" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.938304 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t8fhv"/"kube-root-ca.crt" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.962817 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bskd9\" (UniqueName: \"kubernetes.io/projected/050e068e-c05a-4115-8a20-381ecb7747c6-kube-api-access-bskd9\") pod \"must-gather-wpn7c\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.963292 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/050e068e-c05a-4115-8a20-381ecb7747c6-must-gather-output\") pod \"must-gather-wpn7c\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:25 crc kubenswrapper[4778]: I0312 15:16:25.965084 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t8fhv/must-gather-wpn7c"] Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.065392 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bskd9\" (UniqueName: \"kubernetes.io/projected/050e068e-c05a-4115-8a20-381ecb7747c6-kube-api-access-bskd9\") pod \"must-gather-wpn7c\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.065444 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/050e068e-c05a-4115-8a20-381ecb7747c6-must-gather-output\") pod \"must-gather-wpn7c\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.065991 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/050e068e-c05a-4115-8a20-381ecb7747c6-must-gather-output\") pod \"must-gather-wpn7c\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.084345 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bskd9\" (UniqueName: \"kubernetes.io/projected/050e068e-c05a-4115-8a20-381ecb7747c6-kube-api-access-bskd9\") pod \"must-gather-wpn7c\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.268530 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.733071 4778 scope.go:117] "RemoveContainer" containerID="66d173277dbb8cde37f4f992e677953055661368f74064cf032011267c61214c" Mar 12 15:16:26 crc kubenswrapper[4778]: I0312 15:16:26.765736 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t8fhv/must-gather-wpn7c"] Mar 12 15:16:27 crc kubenswrapper[4778]: I0312 15:16:27.406754 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" event={"ID":"050e068e-c05a-4115-8a20-381ecb7747c6","Type":"ContainerStarted","Data":"91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346"} Mar 12 15:16:27 crc kubenswrapper[4778]: I0312 15:16:27.406803 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" event={"ID":"050e068e-c05a-4115-8a20-381ecb7747c6","Type":"ContainerStarted","Data":"c264bbda6eab92a22196a6070c2ebd25ad32a4313adb11844aceb3d6d0e21de8"} Mar 12 15:16:28 crc kubenswrapper[4778]: I0312 15:16:28.421827 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" event={"ID":"050e068e-c05a-4115-8a20-381ecb7747c6","Type":"ContainerStarted","Data":"115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3"} Mar 12 15:16:28 crc kubenswrapper[4778]: I0312 15:16:28.456566 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" podStartSLOduration=3.456533814 podStartE2EDuration="3.456533814s" podCreationTimestamp="2026-03-12 15:16:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 15:16:28.440823247 +0000 UTC m=+7606.889518703" watchObservedRunningTime="2026-03-12 15:16:28.456533814 +0000 UTC m=+7606.905229250" Mar 12 15:16:28 crc kubenswrapper[4778]: I0312 15:16:28.558213 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:16:28 crc kubenswrapper[4778]: I0312 15:16:28.558280 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.383018 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-vvl5p"] Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.385992 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.388585 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t8fhv"/"default-dockercfg-h8m7h" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.483095 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc4sj\" (UniqueName: \"kubernetes.io/projected/183372ad-b2f3-4b05-af1e-96ce5b752768-kube-api-access-fc4sj\") pod \"crc-debug-vvl5p\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.484521 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/183372ad-b2f3-4b05-af1e-96ce5b752768-host\") pod \"crc-debug-vvl5p\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.586748 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc4sj\" (UniqueName: \"kubernetes.io/projected/183372ad-b2f3-4b05-af1e-96ce5b752768-kube-api-access-fc4sj\") pod \"crc-debug-vvl5p\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.587102 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/183372ad-b2f3-4b05-af1e-96ce5b752768-host\") pod \"crc-debug-vvl5p\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.587421 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/183372ad-b2f3-4b05-af1e-96ce5b752768-host\") pod \"crc-debug-vvl5p\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.606805 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc4sj\" (UniqueName: \"kubernetes.io/projected/183372ad-b2f3-4b05-af1e-96ce5b752768-kube-api-access-fc4sj\") pod \"crc-debug-vvl5p\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: I0312 15:16:31.706065 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:16:31 crc kubenswrapper[4778]: W0312 15:16:31.738108 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod183372ad_b2f3_4b05_af1e_96ce5b752768.slice/crio-9ef0dab72eff1169e4868a63381157ceb8e771a86c14fab3398ea56779a2caf3 WatchSource:0}: Error finding container 9ef0dab72eff1169e4868a63381157ceb8e771a86c14fab3398ea56779a2caf3: Status 404 returned error can't find the container with id 9ef0dab72eff1169e4868a63381157ceb8e771a86c14fab3398ea56779a2caf3 Mar 12 15:16:32 crc kubenswrapper[4778]: I0312 15:16:32.453927 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" event={"ID":"183372ad-b2f3-4b05-af1e-96ce5b752768","Type":"ContainerStarted","Data":"88377826de2b8da64a9c4f2434a9dbd5ccc7cbe21ec7e843bd98b2b4d5469479"} Mar 12 15:16:32 crc kubenswrapper[4778]: I0312 15:16:32.454648 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" event={"ID":"183372ad-b2f3-4b05-af1e-96ce5b752768","Type":"ContainerStarted","Data":"9ef0dab72eff1169e4868a63381157ceb8e771a86c14fab3398ea56779a2caf3"} Mar 12 15:16:32 crc kubenswrapper[4778]: I0312 15:16:32.470319 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" podStartSLOduration=1.470300583 podStartE2EDuration="1.470300583s" podCreationTimestamp="2026-03-12 15:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 15:16:32.465911309 +0000 UTC m=+7610.914606705" watchObservedRunningTime="2026-03-12 15:16:32.470300583 +0000 UTC m=+7610.918995979" Mar 12 15:16:58 crc kubenswrapper[4778]: I0312 15:16:58.558344 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:16:58 crc kubenswrapper[4778]: I0312 15:16:58.558908 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:16:58 crc kubenswrapper[4778]: I0312 15:16:58.558965 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 15:16:58 crc kubenswrapper[4778]: I0312 15:16:58.559802 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c0ffa691d48b1023164222bd8c69a88e4e7a89d268ba03833dc6ae4ab4b44b3"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 15:16:58 crc kubenswrapper[4778]: I0312 15:16:58.559857 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://9c0ffa691d48b1023164222bd8c69a88e4e7a89d268ba03833dc6ae4ab4b44b3" gracePeriod=600 Mar 12 15:16:59 crc kubenswrapper[4778]: I0312 15:16:59.691659 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="9c0ffa691d48b1023164222bd8c69a88e4e7a89d268ba03833dc6ae4ab4b44b3" exitCode=0 Mar 12 15:16:59 crc kubenswrapper[4778]: I0312 15:16:59.691739 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"9c0ffa691d48b1023164222bd8c69a88e4e7a89d268ba03833dc6ae4ab4b44b3"} Mar 12 15:16:59 crc kubenswrapper[4778]: I0312 15:16:59.692373 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc"} Mar 12 15:16:59 crc kubenswrapper[4778]: I0312 15:16:59.692403 4778 scope.go:117] "RemoveContainer" containerID="0d1560644663063f80ae67feb786777f7400aa5bf8ea2f2418887c809789d930" Mar 12 15:17:14 crc kubenswrapper[4778]: I0312 15:17:14.846534 4778 generic.go:334] "Generic (PLEG): container finished" podID="183372ad-b2f3-4b05-af1e-96ce5b752768" containerID="88377826de2b8da64a9c4f2434a9dbd5ccc7cbe21ec7e843bd98b2b4d5469479" exitCode=0 Mar 12 15:17:14 crc kubenswrapper[4778]: I0312 15:17:14.846664 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" event={"ID":"183372ad-b2f3-4b05-af1e-96ce5b752768","Type":"ContainerDied","Data":"88377826de2b8da64a9c4f2434a9dbd5ccc7cbe21ec7e843bd98b2b4d5469479"} Mar 12 15:17:15 crc kubenswrapper[4778]: I0312 15:17:15.974373 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.013845 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-vvl5p"] Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.020765 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-vvl5p"] Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.164103 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc4sj\" (UniqueName: \"kubernetes.io/projected/183372ad-b2f3-4b05-af1e-96ce5b752768-kube-api-access-fc4sj\") pod \"183372ad-b2f3-4b05-af1e-96ce5b752768\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.164243 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/183372ad-b2f3-4b05-af1e-96ce5b752768-host\") pod \"183372ad-b2f3-4b05-af1e-96ce5b752768\" (UID: \"183372ad-b2f3-4b05-af1e-96ce5b752768\") " Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.164340 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/183372ad-b2f3-4b05-af1e-96ce5b752768-host" (OuterVolumeSpecName: "host") pod "183372ad-b2f3-4b05-af1e-96ce5b752768" (UID: "183372ad-b2f3-4b05-af1e-96ce5b752768"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.164713 4778 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/183372ad-b2f3-4b05-af1e-96ce5b752768-host\") on node \"crc\" DevicePath \"\"" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.169573 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/183372ad-b2f3-4b05-af1e-96ce5b752768-kube-api-access-fc4sj" (OuterVolumeSpecName: "kube-api-access-fc4sj") pod "183372ad-b2f3-4b05-af1e-96ce5b752768" (UID: "183372ad-b2f3-4b05-af1e-96ce5b752768"). InnerVolumeSpecName "kube-api-access-fc4sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.264863 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="183372ad-b2f3-4b05-af1e-96ce5b752768" path="/var/lib/kubelet/pods/183372ad-b2f3-4b05-af1e-96ce5b752768/volumes" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.267935 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc4sj\" (UniqueName: \"kubernetes.io/projected/183372ad-b2f3-4b05-af1e-96ce5b752768-kube-api-access-fc4sj\") on node \"crc\" DevicePath \"\"" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.869417 4778 scope.go:117] "RemoveContainer" containerID="88377826de2b8da64a9c4f2434a9dbd5ccc7cbe21ec7e843bd98b2b4d5469479" Mar 12 15:17:16 crc kubenswrapper[4778]: I0312 15:17:16.869655 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-vvl5p" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.316001 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-t6zbj"] Mar 12 15:17:17 crc kubenswrapper[4778]: E0312 15:17:17.316794 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="183372ad-b2f3-4b05-af1e-96ce5b752768" containerName="container-00" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.316813 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="183372ad-b2f3-4b05-af1e-96ce5b752768" containerName="container-00" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.317070 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="183372ad-b2f3-4b05-af1e-96ce5b752768" containerName="container-00" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.317908 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.321734 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t8fhv"/"default-dockercfg-h8m7h" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.407021 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdk54\" (UniqueName: \"kubernetes.io/projected/709dd41c-fb43-4c5d-9741-f407d99cf786-kube-api-access-hdk54\") pod \"crc-debug-t6zbj\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.407212 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/709dd41c-fb43-4c5d-9741-f407d99cf786-host\") pod \"crc-debug-t6zbj\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.509440 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdk54\" (UniqueName: \"kubernetes.io/projected/709dd41c-fb43-4c5d-9741-f407d99cf786-kube-api-access-hdk54\") pod \"crc-debug-t6zbj\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.509544 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/709dd41c-fb43-4c5d-9741-f407d99cf786-host\") pod \"crc-debug-t6zbj\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.509651 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/709dd41c-fb43-4c5d-9741-f407d99cf786-host\") pod \"crc-debug-t6zbj\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.530880 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdk54\" (UniqueName: \"kubernetes.io/projected/709dd41c-fb43-4c5d-9741-f407d99cf786-kube-api-access-hdk54\") pod \"crc-debug-t6zbj\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.633232 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:17 crc kubenswrapper[4778]: W0312 15:17:17.665308 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod709dd41c_fb43_4c5d_9741_f407d99cf786.slice/crio-eeb34a6e12cb7caf13d724302b91bb072707483f0abcda7540aefe215fbdaa58 WatchSource:0}: Error finding container eeb34a6e12cb7caf13d724302b91bb072707483f0abcda7540aefe215fbdaa58: Status 404 returned error can't find the container with id eeb34a6e12cb7caf13d724302b91bb072707483f0abcda7540aefe215fbdaa58 Mar 12 15:17:17 crc kubenswrapper[4778]: I0312 15:17:17.881218 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" event={"ID":"709dd41c-fb43-4c5d-9741-f407d99cf786","Type":"ContainerStarted","Data":"eeb34a6e12cb7caf13d724302b91bb072707483f0abcda7540aefe215fbdaa58"} Mar 12 15:17:18 crc kubenswrapper[4778]: I0312 15:17:18.894242 4778 generic.go:334] "Generic (PLEG): container finished" podID="709dd41c-fb43-4c5d-9741-f407d99cf786" containerID="bbbdcd9b6771dbb5cbdf7ae29f037ff7ca335ae97a393f2a8c10c715cd6d06ac" exitCode=0 Mar 12 15:17:18 crc kubenswrapper[4778]: I0312 15:17:18.894346 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" event={"ID":"709dd41c-fb43-4c5d-9741-f407d99cf786","Type":"ContainerDied","Data":"bbbdcd9b6771dbb5cbdf7ae29f037ff7ca335ae97a393f2a8c10c715cd6d06ac"} Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.023437 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.061927 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdk54\" (UniqueName: \"kubernetes.io/projected/709dd41c-fb43-4c5d-9741-f407d99cf786-kube-api-access-hdk54\") pod \"709dd41c-fb43-4c5d-9741-f407d99cf786\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.062247 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/709dd41c-fb43-4c5d-9741-f407d99cf786-host\") pod \"709dd41c-fb43-4c5d-9741-f407d99cf786\" (UID: \"709dd41c-fb43-4c5d-9741-f407d99cf786\") " Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.062622 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/709dd41c-fb43-4c5d-9741-f407d99cf786-host" (OuterVolumeSpecName: "host") pod "709dd41c-fb43-4c5d-9741-f407d99cf786" (UID: "709dd41c-fb43-4c5d-9741-f407d99cf786"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.077256 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709dd41c-fb43-4c5d-9741-f407d99cf786-kube-api-access-hdk54" (OuterVolumeSpecName: "kube-api-access-hdk54") pod "709dd41c-fb43-4c5d-9741-f407d99cf786" (UID: "709dd41c-fb43-4c5d-9741-f407d99cf786"). InnerVolumeSpecName "kube-api-access-hdk54". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.164357 4778 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/709dd41c-fb43-4c5d-9741-f407d99cf786-host\") on node \"crc\" DevicePath \"\"" Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.164382 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdk54\" (UniqueName: \"kubernetes.io/projected/709dd41c-fb43-4c5d-9741-f407d99cf786-kube-api-access-hdk54\") on node \"crc\" DevicePath \"\"" Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.930301 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" event={"ID":"709dd41c-fb43-4c5d-9741-f407d99cf786","Type":"ContainerDied","Data":"eeb34a6e12cb7caf13d724302b91bb072707483f0abcda7540aefe215fbdaa58"} Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.930623 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeb34a6e12cb7caf13d724302b91bb072707483f0abcda7540aefe215fbdaa58" Mar 12 15:17:20 crc kubenswrapper[4778]: I0312 15:17:20.930369 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-t6zbj" Mar 12 15:17:21 crc kubenswrapper[4778]: I0312 15:17:21.239070 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-t6zbj"] Mar 12 15:17:21 crc kubenswrapper[4778]: I0312 15:17:21.248329 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-t6zbj"] Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.263931 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709dd41c-fb43-4c5d-9741-f407d99cf786" path="/var/lib/kubelet/pods/709dd41c-fb43-4c5d-9741-f407d99cf786/volumes" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.465087 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-sf9lm"] Mar 12 15:17:22 crc kubenswrapper[4778]: E0312 15:17:22.465822 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709dd41c-fb43-4c5d-9741-f407d99cf786" containerName="container-00" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.465843 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="709dd41c-fb43-4c5d-9741-f407d99cf786" containerName="container-00" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.466068 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="709dd41c-fb43-4c5d-9741-f407d99cf786" containerName="container-00" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.466832 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.469478 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t8fhv"/"default-dockercfg-h8m7h" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.506849 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdgwn\" (UniqueName: \"kubernetes.io/projected/d48fd056-e0ab-4645-a4ec-af8315c4c789-kube-api-access-bdgwn\") pod \"crc-debug-sf9lm\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.506943 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d48fd056-e0ab-4645-a4ec-af8315c4c789-host\") pod \"crc-debug-sf9lm\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.608116 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdgwn\" (UniqueName: \"kubernetes.io/projected/d48fd056-e0ab-4645-a4ec-af8315c4c789-kube-api-access-bdgwn\") pod \"crc-debug-sf9lm\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.608174 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d48fd056-e0ab-4645-a4ec-af8315c4c789-host\") pod \"crc-debug-sf9lm\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.608337 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d48fd056-e0ab-4645-a4ec-af8315c4c789-host\") pod \"crc-debug-sf9lm\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.625110 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdgwn\" (UniqueName: \"kubernetes.io/projected/d48fd056-e0ab-4645-a4ec-af8315c4c789-kube-api-access-bdgwn\") pod \"crc-debug-sf9lm\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.865789 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:22 crc kubenswrapper[4778]: I0312 15:17:22.955419 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" event={"ID":"d48fd056-e0ab-4645-a4ec-af8315c4c789","Type":"ContainerStarted","Data":"de2759ba12e8a1c4c30b2c75fa5e913ba606536a3d7f9921e46bf2c30dd4dc93"} Mar 12 15:17:23 crc kubenswrapper[4778]: I0312 15:17:23.964452 4778 generic.go:334] "Generic (PLEG): container finished" podID="d48fd056-e0ab-4645-a4ec-af8315c4c789" containerID="978012010ad73303d32aa324be6cdc810360907cfeab7854465081f99f04817c" exitCode=0 Mar 12 15:17:23 crc kubenswrapper[4778]: I0312 15:17:23.964522 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" event={"ID":"d48fd056-e0ab-4645-a4ec-af8315c4c789","Type":"ContainerDied","Data":"978012010ad73303d32aa324be6cdc810360907cfeab7854465081f99f04817c"} Mar 12 15:17:24 crc kubenswrapper[4778]: I0312 15:17:24.004465 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-sf9lm"] Mar 12 15:17:24 crc kubenswrapper[4778]: I0312 15:17:24.014813 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t8fhv/crc-debug-sf9lm"] Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.072342 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.213928 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdgwn\" (UniqueName: \"kubernetes.io/projected/d48fd056-e0ab-4645-a4ec-af8315c4c789-kube-api-access-bdgwn\") pod \"d48fd056-e0ab-4645-a4ec-af8315c4c789\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.214064 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d48fd056-e0ab-4645-a4ec-af8315c4c789-host\") pod \"d48fd056-e0ab-4645-a4ec-af8315c4c789\" (UID: \"d48fd056-e0ab-4645-a4ec-af8315c4c789\") " Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.214147 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d48fd056-e0ab-4645-a4ec-af8315c4c789-host" (OuterVolumeSpecName: "host") pod "d48fd056-e0ab-4645-a4ec-af8315c4c789" (UID: "d48fd056-e0ab-4645-a4ec-af8315c4c789"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.215003 4778 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d48fd056-e0ab-4645-a4ec-af8315c4c789-host\") on node \"crc\" DevicePath \"\"" Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.221516 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d48fd056-e0ab-4645-a4ec-af8315c4c789-kube-api-access-bdgwn" (OuterVolumeSpecName: "kube-api-access-bdgwn") pod "d48fd056-e0ab-4645-a4ec-af8315c4c789" (UID: "d48fd056-e0ab-4645-a4ec-af8315c4c789"). InnerVolumeSpecName "kube-api-access-bdgwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.318172 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdgwn\" (UniqueName: \"kubernetes.io/projected/d48fd056-e0ab-4645-a4ec-af8315c4c789-kube-api-access-bdgwn\") on node \"crc\" DevicePath \"\"" Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.986161 4778 scope.go:117] "RemoveContainer" containerID="978012010ad73303d32aa324be6cdc810360907cfeab7854465081f99f04817c" Mar 12 15:17:25 crc kubenswrapper[4778]: I0312 15:17:25.986346 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/crc-debug-sf9lm" Mar 12 15:17:26 crc kubenswrapper[4778]: I0312 15:17:26.266277 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d48fd056-e0ab-4645-a4ec-af8315c4c789" path="/var/lib/kubelet/pods/d48fd056-e0ab-4645-a4ec-af8315c4c789/volumes" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.167889 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555478-fcjb2"] Mar 12 15:18:00 crc kubenswrapper[4778]: E0312 15:18:00.169093 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48fd056-e0ab-4645-a4ec-af8315c4c789" containerName="container-00" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.169115 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48fd056-e0ab-4645-a4ec-af8315c4c789" containerName="container-00" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.169514 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48fd056-e0ab-4645-a4ec-af8315c4c789" containerName="container-00" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.170565 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.175352 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.175445 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.175603 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.186548 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555478-fcjb2"] Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.243586 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggbz\" (UniqueName: \"kubernetes.io/projected/341f8c65-027d-48d0-b0c2-b843867c2413-kube-api-access-2ggbz\") pod \"auto-csr-approver-29555478-fcjb2\" (UID: \"341f8c65-027d-48d0-b0c2-b843867c2413\") " pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.344990 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggbz\" (UniqueName: \"kubernetes.io/projected/341f8c65-027d-48d0-b0c2-b843867c2413-kube-api-access-2ggbz\") pod \"auto-csr-approver-29555478-fcjb2\" (UID: \"341f8c65-027d-48d0-b0c2-b843867c2413\") " pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.372030 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggbz\" (UniqueName: \"kubernetes.io/projected/341f8c65-027d-48d0-b0c2-b843867c2413-kube-api-access-2ggbz\") pod \"auto-csr-approver-29555478-fcjb2\" (UID: \"341f8c65-027d-48d0-b0c2-b843867c2413\") " pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.496268 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:00 crc kubenswrapper[4778]: I0312 15:18:00.959460 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555478-fcjb2"] Mar 12 15:18:01 crc kubenswrapper[4778]: I0312 15:18:01.306472 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" event={"ID":"341f8c65-027d-48d0-b0c2-b843867c2413","Type":"ContainerStarted","Data":"dac606d016de3543c499198eb7ee631d3101d78cca57ba65c393a13f9eceedb7"} Mar 12 15:18:01 crc kubenswrapper[4778]: I0312 15:18:01.727156 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-86cb765474-5pq5z_6bd172c5-383f-4273-98a5-2c92223dc765/barbican-api/0.log" Mar 12 15:18:01 crc kubenswrapper[4778]: I0312 15:18:01.867009 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-86cb765474-5pq5z_6bd172c5-383f-4273-98a5-2c92223dc765/barbican-api-log/0.log" Mar 12 15:18:01 crc kubenswrapper[4778]: I0312 15:18:01.917279 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65c9994dfd-xznqh_8ee1f546-8428-4b23-93e4-b8370fd4224b/barbican-keystone-listener/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.207147 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-65c9994dfd-xznqh_8ee1f546-8428-4b23-93e4-b8370fd4224b/barbican-keystone-listener-log/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.222305 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7dcf9787-ngc87_d505bb59-3c9e-4cfa-891c-c8e0068e2567/barbican-worker/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.234758 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7dcf9787-ngc87_d505bb59-3c9e-4cfa-891c-c8e0068e2567/barbican-worker-log/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.534230 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ntpnx_b99627a8-43d8-4f7d-90f7-530eda3c2213/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.549104 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/ceilometer-central-agent/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.715809 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/proxy-httpd/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.735503 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/sg-core/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.919898 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9f1d0355-a73a-4a93-94fb-b439436cf1b1/ceilometer-notification-agent/0.log" Mar 12 15:18:02 crc kubenswrapper[4778]: I0312 15:18:02.989698 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_99f72014-50e8-4dd4-9764-1b2c7d546b30/cinder-api-log/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.066743 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_99f72014-50e8-4dd4-9764-1b2c7d546b30/cinder-api/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.154415 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_39ee2404-53a8-4598-8c4b-c3a34fbf3480/cinder-scheduler/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.257591 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_39ee2404-53a8-4598-8c4b-c3a34fbf3480/probe/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.343416 4778 generic.go:334] "Generic (PLEG): container finished" podID="341f8c65-027d-48d0-b0c2-b843867c2413" containerID="27e746629157759d4e60a414cb672470c7ab54258b384fb1bc8e845de836c293" exitCode=0 Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.343464 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" event={"ID":"341f8c65-027d-48d0-b0c2-b843867c2413","Type":"ContainerDied","Data":"27e746629157759d4e60a414cb672470c7ab54258b384fb1bc8e845de836c293"} Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.359086 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4szjl_5c5541f3-fb44-476b-91c2-b07dffe50894/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.473528 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jg9z6_36bb4acd-fab3-4998-a8cd-a6ebcc800fc8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.605289 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/init/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.763458 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/init/0.log" Mar 12 15:18:03 crc kubenswrapper[4778]: I0312 15:18:03.829319 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-2xksx_96ba9a1b-ae5f-4b42-b8eb-1f0e3656ae61/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:04 crc kubenswrapper[4778]: I0312 15:18:04.083825 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_81c1a05c-5642-43d4-8a7b-229330168332/glance-log/0.log" Mar 12 15:18:04 crc kubenswrapper[4778]: I0312 15:18:04.119618 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_81c1a05c-5642-43d4-8a7b-229330168332/glance-httpd/0.log" Mar 12 15:18:04 crc kubenswrapper[4778]: I0312 15:18:04.210239 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f89cfcd7f-vk6h4_46f34397-57fe-425d-b69d-040f4384ac69/dnsmasq-dns/0.log" Mar 12 15:18:04 crc kubenswrapper[4778]: I0312 15:18:04.353637 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7fa757af-1c91-4b93-8916-5bbd99b8522e/glance-httpd/0.log" Mar 12 15:18:04 crc kubenswrapper[4778]: I0312 15:18:04.367093 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7fa757af-1c91-4b93-8916-5bbd99b8522e/glance-log/0.log" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:04.544819 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bngcx_f69e6cfe-f7c2-4127-b4df-710725c52227/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:04.671623 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g252n_29f8609b-4a3b-42ba-9450-a2b633bb4c2c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:04.694380 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:04.847814 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ggbz\" (UniqueName: \"kubernetes.io/projected/341f8c65-027d-48d0-b0c2-b843867c2413-kube-api-access-2ggbz\") pod \"341f8c65-027d-48d0-b0c2-b843867c2413\" (UID: \"341f8c65-027d-48d0-b0c2-b843867c2413\") " Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:04.856317 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341f8c65-027d-48d0-b0c2-b843867c2413-kube-api-access-2ggbz" (OuterVolumeSpecName: "kube-api-access-2ggbz") pod "341f8c65-027d-48d0-b0c2-b843867c2413" (UID: "341f8c65-027d-48d0-b0c2-b843867c2413"). InnerVolumeSpecName "kube-api-access-2ggbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:04.952198 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ggbz\" (UniqueName: \"kubernetes.io/projected/341f8c65-027d-48d0-b0c2-b843867c2413-kube-api-access-2ggbz\") on node \"crc\" DevicePath \"\"" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.089333 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555401-vjgkl_e4df6927-3452-4b36-b59a-a1fdcd4272a4/keystone-cron/0.log" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.316351 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29555461-lmqk9_ebdf3274-70cb-4083-bf12-5d1038a9b7ba/keystone-cron/0.log" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.359866 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" event={"ID":"341f8c65-027d-48d0-b0c2-b843867c2413","Type":"ContainerDied","Data":"dac606d016de3543c499198eb7ee631d3101d78cca57ba65c393a13f9eceedb7"} Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.359913 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dac606d016de3543c499198eb7ee631d3101d78cca57ba65c393a13f9eceedb7" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.359931 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555478-fcjb2" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.507045 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_51f24fcd-aff5-4785-abf7-4936180cee78/kube-state-metrics/0.log" Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.763703 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555472-k4f9h"] Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.783585 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555472-k4f9h"] Mar 12 15:18:05 crc kubenswrapper[4778]: I0312 15:18:05.900803 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4m9w8_8713b951-b516-42bd-9286-4343e5bcc955/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:06 crc kubenswrapper[4778]: I0312 15:18:06.175494 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-6lrlq_a56bb599-f10d-4564-b6bf-48128dc2c7f1/keystone-api/0.log" Mar 12 15:18:06 crc kubenswrapper[4778]: I0312 15:18:06.263150 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30c4e913-d163-4764-8738-ac336cd93df9" path="/var/lib/kubelet/pods/30c4e913-d163-4764-8738-ac336cd93df9/volumes" Mar 12 15:18:06 crc kubenswrapper[4778]: I0312 15:18:06.315917 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69b6dc4885-z4h9m_16dea17b-eaa4-4bbf-8895-c077b3e28d66/keystone-api/0.log" Mar 12 15:18:07 crc kubenswrapper[4778]: I0312 15:18:07.056003 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-dggmh_7596a69e-33c9-4a2b-89fc-e4c41252b3fd/neutron-httpd/0.log" Mar 12 15:18:07 crc kubenswrapper[4778]: I0312 15:18:07.294338 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-zx97x_8a67d4b7-d8eb-40f4-b51d-62e92c6042c1/neutron-httpd/0.log" Mar 12 15:18:07 crc kubenswrapper[4778]: I0312 15:18:07.483012 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-custom-edpm-deployment-openstack-edpm-ipawlfsg_5cc410de-5b42-44d1-8b29-37161475730e/neutron-metadata-custom-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:09 crc kubenswrapper[4778]: I0312 15:18:09.549047 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ec63cc68-6fde-419b-973c-91fc982e6a49/memcached/0.log" Mar 12 15:18:10 crc kubenswrapper[4778]: I0312 15:18:10.279583 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4/nova-api-log/0.log" Mar 12 15:18:11 crc kubenswrapper[4778]: I0312 15:18:11.689409 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-zx97x_8a67d4b7-d8eb-40f4-b51d-62e92c6042c1/neutron-api/0.log" Mar 12 15:18:12 crc kubenswrapper[4778]: I0312 15:18:12.491856 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-1_f0341d80-4327-4c9e-bc11-0cddbc6eab66/nova-api-log/0.log" Mar 12 15:18:12 crc kubenswrapper[4778]: I0312 15:18:12.958263 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_929bb450-949d-4f4f-9c21-de6c3fe32927/nova-cell0-conductor-conductor/0.log" Mar 12 15:18:13 crc kubenswrapper[4778]: I0312 15:18:13.277368 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1466aea3-fa10-49a6-a254-a96a52091aca/nova-cell1-conductor-conductor/0.log" Mar 12 15:18:13 crc kubenswrapper[4778]: I0312 15:18:13.369945 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_13b8e1df-5a8c-44de-b8e8-6c7efdb8bad4/nova-api-api/0.log" Mar 12 15:18:13 crc kubenswrapper[4778]: I0312 15:18:13.462260 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-566c4d5fc-dggmh_7596a69e-33c9-4a2b-89fc-e4c41252b3fd/neutron-api/0.log" Mar 12 15:18:13 crc kubenswrapper[4778]: I0312 15:18:13.536213 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-metadata-0_c289a520-78eb-433f-b7a4-0c03be917c18/nova-cell1-metadata-log/0.log" Mar 12 15:18:13 crc kubenswrapper[4778]: I0312 15:18:13.918925 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2b43a8b1-b8bc-4ab5-af66-674fa7ff47d7/nova-cell1-novncproxy-novncproxy/0.log" Mar 12 15:18:13 crc kubenswrapper[4778]: I0312 15:18:13.943386 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5tw6s_6ed77f87-e6b2-4c7a-8b0e-003106200dc8/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.254140 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/mysql-bootstrap/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.409663 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/mysql-bootstrap/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.473081 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fe52f8ba-9053-4733-b2e3-8f1becf437c8/galera/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.618158 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-metadata-0_c289a520-78eb-433f-b7a4-0c03be917c18/nova-cell1-metadata-metadata/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.724737 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/mysql-bootstrap/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.771680 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-1_f0341d80-4327-4c9e-bc11-0cddbc6eab66/nova-api-api/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.973260 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/galera/0.log" Mar 12 15:18:14 crc kubenswrapper[4778]: I0312 15:18:14.998656 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_856cd6d1-db21-4503-94d7-cbf27ca96cc2/openstackclient/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.001874 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_663feb48-0ed1-4947-97c3-e0bac206fdb2/mysql-bootstrap/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.158318 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f613745b-fe33-4918-9e0a-da2a59c55e33/nova-scheduler-scheduler/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.179754 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vtt4z_a8484e5d-6f77-407c-81db-0d9b2a6b37fd/openstack-network-exporter/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.191121 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4wct6_3b8efd1e-884d-4963-b69f-04ede0a92267/ovn-controller/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.403325 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server-init/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.516769 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server-init/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.547092 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovsdb-server/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.556572 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p67vh_bd159b65-0c66-4809-949e-0f1babbaa8e6/ovs-vswitchd/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.658941 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9lbdq_3c0a2200-506d-4ac3-b08c-9b3156c9e573/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.733621 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b25f9c9-784a-4a52-9bb3-02c6c4592702/openstack-network-exporter/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.760526 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b25f9c9-784a-4a52-9bb3-02c6c4592702/ovn-northd/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.870798 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7321e15e-673c-4e0d-80f8-6ac644c1940f/openstack-network-exporter/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.922795 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7321e15e-673c-4e0d-80f8-6ac644c1940f/ovsdbserver-nb/0.log" Mar 12 15:18:15 crc kubenswrapper[4778]: I0312 15:18:15.948248 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c951c6f-06fd-4793-a95b-26b5c1400d73/openstack-network-exporter/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.065681 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c951c6f-06fd-4793-a95b-26b5c1400d73/ovsdbserver-sb/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.230588 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/setup-container/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.539590 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/setup-container/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.542133 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_629c84c5-e6cf-4aa7-ba9a-5a5fe7f53a03/rabbitmq/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.562978 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d4d765698-l7bjx_267e7df2-d35c-45c4-af65-e8af31f8f6cf/placement-api/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.705963 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/setup-container/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.722373 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d4d765698-l7bjx_267e7df2-d35c-45c4-af65-e8af31f8f6cf/placement-log/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.894445 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/setup-container/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.963198 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wcdkc_43a3ffe4-8b64-4e26-b63a-5254a986e4a4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:16 crc kubenswrapper[4778]: I0312 15:18:16.987999 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1e89dfcc-2ac3-444c-91e8-56991eae096b/rabbitmq/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.099378 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6nfzc_bd7ac6b4-5600-45ce-b0ea-199dd4baefcb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.188483 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gt58t_b0bb06df-44bb-4939-9492-a6ad3d6b5368/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.221667 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8mmjm_c993b33e-6c36-4524-864a-65da461a8e0c/ssh-known-hosts-edpm-deployment/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.461172 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-77f887c49f-fw2qd_bbd76cb8-462f-4e60-b755-ef3170e70d11/proxy-server/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.552510 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5knbg_2edc2c90-f91e-402d-809c-514e9d8a5e04/swift-ring-rebalance/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.671639 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-77f887c49f-fw2qd_bbd76cb8-462f-4e60-b755-ef3170e70d11/proxy-httpd/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.680468 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-auditor/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.691430 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-reaper/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.775822 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-replicator/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.855236 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-auditor/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.870381 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/account-server/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.934903 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-replicator/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.944820 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-server/0.log" Mar 12 15:18:17 crc kubenswrapper[4778]: I0312 15:18:17.993356 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/container-updater/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.089716 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-auditor/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.112357 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-expirer/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.170824 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-replicator/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.176219 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-server/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.198681 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/object-updater/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.293704 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/rsync/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.302630 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c01f943c-e09c-4727-8cf7-eec58a56b363/swift-recon-cron/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.432893 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qrk5s_2bfaafaf-36fb-4f1a-99ed-abb8b7bb4ae1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.555279 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_74897d0a-ca7b-4589-bd4c-75910c2d491c/tempest-tests-tempest-tests-runner/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.566997 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_82246f69-2112-44e9-a783-a4a5926188b4/test-operator-logs-container/0.log" Mar 12 15:18:18 crc kubenswrapper[4778]: I0312 15:18:18.651928 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9glvr_41583476-38cd-4c0d-a05a-96ddc5b330ca/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 15:18:26 crc kubenswrapper[4778]: I0312 15:18:26.867131 4778 scope.go:117] "RemoveContainer" containerID="971c448e63690dd43ac1d65335a70f73b2547d4337b42531c9336354c82b33f3" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.289576 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zmsmk"] Mar 12 15:18:39 crc kubenswrapper[4778]: E0312 15:18:39.292986 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341f8c65-027d-48d0-b0c2-b843867c2413" containerName="oc" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.293145 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="341f8c65-027d-48d0-b0c2-b843867c2413" containerName="oc" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.293561 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="341f8c65-027d-48d0-b0c2-b843867c2413" containerName="oc" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.295731 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.330250 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zmsmk"] Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.471522 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-utilities\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.471887 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-catalog-content\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.472041 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbmb8\" (UniqueName: \"kubernetes.io/projected/d07c4f66-dce0-41f8-8978-a88beb6bead5-kube-api-access-wbmb8\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.573765 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbmb8\" (UniqueName: \"kubernetes.io/projected/d07c4f66-dce0-41f8-8978-a88beb6bead5-kube-api-access-wbmb8\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.574736 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-utilities\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.575331 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-catalog-content\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.575240 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-utilities\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.575629 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-catalog-content\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.597805 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbmb8\" (UniqueName: \"kubernetes.io/projected/d07c4f66-dce0-41f8-8978-a88beb6bead5-kube-api-access-wbmb8\") pod \"certified-operators-zmsmk\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:39 crc kubenswrapper[4778]: I0312 15:18:39.618909 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:40 crc kubenswrapper[4778]: I0312 15:18:40.152974 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zmsmk"] Mar 12 15:18:40 crc kubenswrapper[4778]: I0312 15:18:40.641577 4778 generic.go:334] "Generic (PLEG): container finished" podID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerID="7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f" exitCode=0 Mar 12 15:18:40 crc kubenswrapper[4778]: I0312 15:18:40.641689 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerDied","Data":"7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f"} Mar 12 15:18:40 crc kubenswrapper[4778]: I0312 15:18:40.641838 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerStarted","Data":"587d5bcfefe3b906ce3068b4e7ec750fe1f4761ad2d6b33acd8486e1cc6df24b"} Mar 12 15:18:41 crc kubenswrapper[4778]: I0312 15:18:41.612004 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/util/0.log" Mar 12 15:18:41 crc kubenswrapper[4778]: I0312 15:18:41.650961 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerStarted","Data":"b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2"} Mar 12 15:18:41 crc kubenswrapper[4778]: I0312 15:18:41.821131 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/pull/0.log" Mar 12 15:18:41 crc kubenswrapper[4778]: I0312 15:18:41.821208 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/util/0.log" Mar 12 15:18:41 crc kubenswrapper[4778]: I0312 15:18:41.822481 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/pull/0.log" Mar 12 15:18:42 crc kubenswrapper[4778]: I0312 15:18:42.018149 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/util/0.log" Mar 12 15:18:42 crc kubenswrapper[4778]: I0312 15:18:42.030692 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/extract/0.log" Mar 12 15:18:42 crc kubenswrapper[4778]: I0312 15:18:42.039147 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d52f25c614d14ea9d555eaa2e62114c0c7d01977d30b495569716fcde5fmfr_e1d0ffee-229e-4da3-ac89-02bf6f6a439f/pull/0.log" Mar 12 15:18:42 crc kubenswrapper[4778]: I0312 15:18:42.444412 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-9n6jv_ad531191-d7c5-4ef6-9929-3a5869751d98/manager/0.log" Mar 12 15:18:42 crc kubenswrapper[4778]: I0312 15:18:42.796433 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-gknp2_db7f6b97-2903-44bf-803f-c00c337400b9/manager/0.log" Mar 12 15:18:43 crc kubenswrapper[4778]: I0312 15:18:43.071007 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-b7tkm_e290c1ea-a39d-451e-a24b-17a2b61ff6f0/manager/0.log" Mar 12 15:18:43 crc kubenswrapper[4778]: I0312 15:18:43.275409 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-4jgt8_4c2bf703-ecc1-4bb1-aa03-a64e55dfdb71/manager/0.log" Mar 12 15:18:43 crc kubenswrapper[4778]: I0312 15:18:43.671290 4778 generic.go:334] "Generic (PLEG): container finished" podID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerID="b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2" exitCode=0 Mar 12 15:18:43 crc kubenswrapper[4778]: I0312 15:18:43.671336 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerDied","Data":"b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2"} Mar 12 15:18:43 crc kubenswrapper[4778]: I0312 15:18:43.853836 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-5d6qz_02bc06ca-f4e6-4fde-bd5d-882714d9652c/manager/0.log" Mar 12 15:18:43 crc kubenswrapper[4778]: I0312 15:18:43.906876 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-qb8s8_98a4cfbd-3037-48b5-9047-5d574dcc0aca/manager/0.log" Mar 12 15:18:44 crc kubenswrapper[4778]: I0312 15:18:44.230755 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7dxdh_7e02c37f-b9af-46c9-a743-03ead9b060db/manager/0.log" Mar 12 15:18:44 crc kubenswrapper[4778]: I0312 15:18:44.465704 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-pn8tk_5e38a4fd-95f8-437b-923b-eca33b1387e6/manager/0.log" Mar 12 15:18:44 crc kubenswrapper[4778]: I0312 15:18:44.681613 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerStarted","Data":"1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d"} Mar 12 15:18:44 crc kubenswrapper[4778]: I0312 15:18:44.710940 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zmsmk" podStartSLOduration=2.220049151 podStartE2EDuration="5.710917236s" podCreationTimestamp="2026-03-12 15:18:39 +0000 UTC" firstStartedPulling="2026-03-12 15:18:40.64417052 +0000 UTC m=+7739.092865916" lastFinishedPulling="2026-03-12 15:18:44.135038605 +0000 UTC m=+7742.583734001" observedRunningTime="2026-03-12 15:18:44.701038625 +0000 UTC m=+7743.149734081" watchObservedRunningTime="2026-03-12 15:18:44.710917236 +0000 UTC m=+7743.159612642" Mar 12 15:18:44 crc kubenswrapper[4778]: I0312 15:18:44.712151 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-jlbft_2d577800-0ee1-4fe5-a7fb-8794fb8c4c6f/manager/0.log" Mar 12 15:18:45 crc kubenswrapper[4778]: I0312 15:18:45.063946 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-dd2ft_076835c9-352b-4e40-80c4-3bce3bb80594/manager/0.log" Mar 12 15:18:45 crc kubenswrapper[4778]: I0312 15:18:45.449949 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-686d5f9fbd-vv9rc_d7288cc6-4247-4d03-bd37-9862243bf613/manager/0.log" Mar 12 15:18:45 crc kubenswrapper[4778]: I0312 15:18:45.584604 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-xm4cc_c8818ac0-af8b-42c9-a923-425fe79ed203/manager/0.log" Mar 12 15:18:45 crc kubenswrapper[4778]: I0312 15:18:45.606373 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-cdgg9_1a01d06c-be6f-45de-a22d-c8f1058a3a84/manager/0.log" Mar 12 15:18:45 crc kubenswrapper[4778]: I0312 15:18:45.903322 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7qq9w6_4f7d316e-6896-4f84-8423-6f79778c1c6b/manager/0.log" Mar 12 15:18:46 crc kubenswrapper[4778]: I0312 15:18:46.052787 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-5bc4df7446-x9bsl_34bbdc16-4518-4ee5-9a70-3cedcc5f0159/operator/0.log" Mar 12 15:18:46 crc kubenswrapper[4778]: I0312 15:18:46.174449 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-b2fsv_748546a6-1355-470f-b8d0-de395cf3f681/registry-server/0.log" Mar 12 15:18:46 crc kubenswrapper[4778]: I0312 15:18:46.463083 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-bbgmb_8d38fd7e-6fa1-4b0c-9c82-9c57290c7837/manager/0.log" Mar 12 15:18:46 crc kubenswrapper[4778]: I0312 15:18:46.584085 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-wvpf8_52524252-25bd-49e5-822e-3d4668aff2f9/manager/0.log" Mar 12 15:18:46 crc kubenswrapper[4778]: I0312 15:18:46.736779 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-shf7b_034f39d8-a33e-4e37-bcde-51fb22debdd1/operator/0.log" Mar 12 15:18:46 crc kubenswrapper[4778]: I0312 15:18:46.943682 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-84mps_64a36384-f2e6-4077-b2ca-de2a6ce6ea06/manager/0.log" Mar 12 15:18:47 crc kubenswrapper[4778]: I0312 15:18:47.088412 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-gfv5z_6ad9bf9f-7214-44bc-a65d-1dcbf385fc2c/manager/0.log" Mar 12 15:18:47 crc kubenswrapper[4778]: I0312 15:18:47.155200 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-pcfrz_ed9b9271-4ae9-440a-9411-15d46267106e/manager/0.log" Mar 12 15:18:47 crc kubenswrapper[4778]: I0312 15:18:47.358551 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-2tjsk_8c02ecb8-0e15-4672-823a-c4437ca5bf8c/manager/0.log" Mar 12 15:18:47 crc kubenswrapper[4778]: I0312 15:18:47.684248 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5785b7957-7vdgw_d0784623-5f08-4109-9c7e-0a329210ce07/manager/0.log" Mar 12 15:18:49 crc kubenswrapper[4778]: I0312 15:18:49.619640 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:49 crc kubenswrapper[4778]: I0312 15:18:49.620650 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:49 crc kubenswrapper[4778]: I0312 15:18:49.672486 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:49 crc kubenswrapper[4778]: I0312 15:18:49.788815 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:49 crc kubenswrapper[4778]: I0312 15:18:49.903684 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zmsmk"] Mar 12 15:18:51 crc kubenswrapper[4778]: I0312 15:18:51.749286 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zmsmk" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="registry-server" containerID="cri-o://1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d" gracePeriod=2 Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.269106 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.402366 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmb8\" (UniqueName: \"kubernetes.io/projected/d07c4f66-dce0-41f8-8978-a88beb6bead5-kube-api-access-wbmb8\") pod \"d07c4f66-dce0-41f8-8978-a88beb6bead5\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.402683 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-utilities\") pod \"d07c4f66-dce0-41f8-8978-a88beb6bead5\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.402789 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-catalog-content\") pod \"d07c4f66-dce0-41f8-8978-a88beb6bead5\" (UID: \"d07c4f66-dce0-41f8-8978-a88beb6bead5\") " Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.404110 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-utilities" (OuterVolumeSpecName: "utilities") pod "d07c4f66-dce0-41f8-8978-a88beb6bead5" (UID: "d07c4f66-dce0-41f8-8978-a88beb6bead5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.419789 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d07c4f66-dce0-41f8-8978-a88beb6bead5-kube-api-access-wbmb8" (OuterVolumeSpecName: "kube-api-access-wbmb8") pod "d07c4f66-dce0-41f8-8978-a88beb6bead5" (UID: "d07c4f66-dce0-41f8-8978-a88beb6bead5"). InnerVolumeSpecName "kube-api-access-wbmb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.492958 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d07c4f66-dce0-41f8-8978-a88beb6bead5" (UID: "d07c4f66-dce0-41f8-8978-a88beb6bead5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.505558 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbmb8\" (UniqueName: \"kubernetes.io/projected/d07c4f66-dce0-41f8-8978-a88beb6bead5-kube-api-access-wbmb8\") on node \"crc\" DevicePath \"\"" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.505586 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.505597 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d07c4f66-dce0-41f8-8978-a88beb6bead5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.759384 4778 generic.go:334] "Generic (PLEG): container finished" podID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerID="1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d" exitCode=0 Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.759424 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerDied","Data":"1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d"} Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.759450 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmsmk" event={"ID":"d07c4f66-dce0-41f8-8978-a88beb6bead5","Type":"ContainerDied","Data":"587d5bcfefe3b906ce3068b4e7ec750fe1f4761ad2d6b33acd8486e1cc6df24b"} Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.759465 4778 scope.go:117] "RemoveContainer" containerID="1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.759600 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmsmk" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.784439 4778 scope.go:117] "RemoveContainer" containerID="b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.795545 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zmsmk"] Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.805035 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zmsmk"] Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.822434 4778 scope.go:117] "RemoveContainer" containerID="7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.864434 4778 scope.go:117] "RemoveContainer" containerID="1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d" Mar 12 15:18:52 crc kubenswrapper[4778]: E0312 15:18:52.865848 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d\": container with ID starting with 1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d not found: ID does not exist" containerID="1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.865890 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d"} err="failed to get container status \"1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d\": rpc error: code = NotFound desc = could not find container \"1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d\": container with ID starting with 1b60f78bfe6e75b52d9353c4adc4bbfd22877bf19966b144cae0b72bfb0cec7d not found: ID does not exist" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.866105 4778 scope.go:117] "RemoveContainer" containerID="b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2" Mar 12 15:18:52 crc kubenswrapper[4778]: E0312 15:18:52.866632 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2\": container with ID starting with b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2 not found: ID does not exist" containerID="b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.866685 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2"} err="failed to get container status \"b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2\": rpc error: code = NotFound desc = could not find container \"b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2\": container with ID starting with b8a4b13d36e704db61600a8494f1afac8230ce7b9ab5570b692d8a13cc92d8a2 not found: ID does not exist" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.866715 4778 scope.go:117] "RemoveContainer" containerID="7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f" Mar 12 15:18:52 crc kubenswrapper[4778]: E0312 15:18:52.867219 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f\": container with ID starting with 7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f not found: ID does not exist" containerID="7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f" Mar 12 15:18:52 crc kubenswrapper[4778]: I0312 15:18:52.867261 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f"} err="failed to get container status \"7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f\": rpc error: code = NotFound desc = could not find container \"7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f\": container with ID starting with 7237711c2c7b10d6207511e0a3ed4c19d27357e3abdfd78772a04042037f685f not found: ID does not exist" Mar 12 15:18:54 crc kubenswrapper[4778]: I0312 15:18:54.262328 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" path="/var/lib/kubelet/pods/d07c4f66-dce0-41f8-8978-a88beb6bead5/volumes" Mar 12 15:18:56 crc kubenswrapper[4778]: I0312 15:18:56.312983 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-6h2c2_ffb8a1f4-4533-4368-a900-95d37fe1d3ad/manager/0.log" Mar 12 15:18:58 crc kubenswrapper[4778]: I0312 15:18:58.557429 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:18:58 crc kubenswrapper[4778]: I0312 15:18:58.557721 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:19:08 crc kubenswrapper[4778]: I0312 15:19:08.090168 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zkrqr_f799c7e9-1c31-40bc-9ece-06a086683a98/control-plane-machine-set-operator/0.log" Mar 12 15:19:08 crc kubenswrapper[4778]: I0312 15:19:08.260606 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-242cb_e2967620-e2ce-4763-8a6c-e5a37f3a1f98/kube-rbac-proxy/0.log" Mar 12 15:19:08 crc kubenswrapper[4778]: I0312 15:19:08.293763 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-242cb_e2967620-e2ce-4763-8a6c-e5a37f3a1f98/machine-api-operator/0.log" Mar 12 15:19:21 crc kubenswrapper[4778]: I0312 15:19:21.019632 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-2774s_92b29110-f478-42b5-9a5f-c9330a3973b2/cert-manager-controller/0.log" Mar 12 15:19:21 crc kubenswrapper[4778]: I0312 15:19:21.205069 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-jxs4g_804d0b09-6fab-4277-936a-5e0324d76b3e/cert-manager-cainjector/0.log" Mar 12 15:19:21 crc kubenswrapper[4778]: I0312 15:19:21.322568 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-ffh2x_45da07c5-bccb-4433-aa38-d9d2894f1b09/cert-manager-webhook/0.log" Mar 12 15:19:28 crc kubenswrapper[4778]: I0312 15:19:28.558000 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:19:28 crc kubenswrapper[4778]: I0312 15:19:28.559446 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:19:33 crc kubenswrapper[4778]: I0312 15:19:33.932314 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-jbxx4_af2d568b-9719-4da9-b0e8-e28d314ed860/nmstate-console-plugin/0.log" Mar 12 15:19:34 crc kubenswrapper[4778]: I0312 15:19:34.121135 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rbsjl_d8309ffe-a26c-44a8-84e2-7b7ec10982a8/nmstate-handler/0.log" Mar 12 15:19:34 crc kubenswrapper[4778]: I0312 15:19:34.146027 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-b2s5h_7855d7b1-c7cf-4b63-9313-051a391fcf43/kube-rbac-proxy/0.log" Mar 12 15:19:34 crc kubenswrapper[4778]: I0312 15:19:34.210262 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-b2s5h_7855d7b1-c7cf-4b63-9313-051a391fcf43/nmstate-metrics/0.log" Mar 12 15:19:34 crc kubenswrapper[4778]: I0312 15:19:34.314360 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-hxzd6_fb85eef5-01f9-4fa6-b9d8-9606d04b8cd3/nmstate-operator/0.log" Mar 12 15:19:34 crc kubenswrapper[4778]: I0312 15:19:34.421205 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-94rbc_ef796a94-b10d-4d18-ae88-f64bc3a6b87d/nmstate-webhook/0.log" Mar 12 15:19:58 crc kubenswrapper[4778]: I0312 15:19:58.557346 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:19:58 crc kubenswrapper[4778]: I0312 15:19:58.557910 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:19:58 crc kubenswrapper[4778]: I0312 15:19:58.557977 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 15:19:58 crc kubenswrapper[4778]: I0312 15:19:58.558887 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 15:19:58 crc kubenswrapper[4778]: I0312 15:19:58.558984 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" gracePeriod=600 Mar 12 15:19:58 crc kubenswrapper[4778]: E0312 15:19:58.686173 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:19:59 crc kubenswrapper[4778]: I0312 15:19:59.369993 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" exitCode=0 Mar 12 15:19:59 crc kubenswrapper[4778]: I0312 15:19:59.370038 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc"} Mar 12 15:19:59 crc kubenswrapper[4778]: I0312 15:19:59.370353 4778 scope.go:117] "RemoveContainer" containerID="9c0ffa691d48b1023164222bd8c69a88e4e7a89d268ba03833dc6ae4ab4b44b3" Mar 12 15:19:59 crc kubenswrapper[4778]: I0312 15:19:59.371157 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:19:59 crc kubenswrapper[4778]: E0312 15:19:59.371478 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.149410 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555480-vwvb5"] Mar 12 15:20:00 crc kubenswrapper[4778]: E0312 15:20:00.149795 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="extract-utilities" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.149808 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="extract-utilities" Mar 12 15:20:00 crc kubenswrapper[4778]: E0312 15:20:00.149834 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="extract-content" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.149840 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="extract-content" Mar 12 15:20:00 crc kubenswrapper[4778]: E0312 15:20:00.149854 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="registry-server" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.149859 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="registry-server" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.150056 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07c4f66-dce0-41f8-8978-a88beb6bead5" containerName="registry-server" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.150694 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.153605 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.157287 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.157738 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.173545 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555480-vwvb5"] Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.262108 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5v8r\" (UniqueName: \"kubernetes.io/projected/9e32f842-16d7-484b-a241-e24ea8d3db45-kube-api-access-f5v8r\") pod \"auto-csr-approver-29555480-vwvb5\" (UID: \"9e32f842-16d7-484b-a241-e24ea8d3db45\") " pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.364554 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5v8r\" (UniqueName: \"kubernetes.io/projected/9e32f842-16d7-484b-a241-e24ea8d3db45-kube-api-access-f5v8r\") pod \"auto-csr-approver-29555480-vwvb5\" (UID: \"9e32f842-16d7-484b-a241-e24ea8d3db45\") " pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.385135 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5v8r\" (UniqueName: \"kubernetes.io/projected/9e32f842-16d7-484b-a241-e24ea8d3db45-kube-api-access-f5v8r\") pod \"auto-csr-approver-29555480-vwvb5\" (UID: \"9e32f842-16d7-484b-a241-e24ea8d3db45\") " pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.469272 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:00 crc kubenswrapper[4778]: I0312 15:20:00.976845 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555480-vwvb5"] Mar 12 15:20:01 crc kubenswrapper[4778]: I0312 15:20:01.402379 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" event={"ID":"9e32f842-16d7-484b-a241-e24ea8d3db45","Type":"ContainerStarted","Data":"e22dd6e038748741ff00db0c5d6a48f31b87069c92a6e6fb50d913d1489dd215"} Mar 12 15:20:03 crc kubenswrapper[4778]: I0312 15:20:03.427679 4778 generic.go:334] "Generic (PLEG): container finished" podID="9e32f842-16d7-484b-a241-e24ea8d3db45" containerID="7d34b9f856d96ac0b056ec93139556664e1582951a8e260f7792f49806a39777" exitCode=0 Mar 12 15:20:03 crc kubenswrapper[4778]: I0312 15:20:03.427776 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" event={"ID":"9e32f842-16d7-484b-a241-e24ea8d3db45","Type":"ContainerDied","Data":"7d34b9f856d96ac0b056ec93139556664e1582951a8e260f7792f49806a39777"} Mar 12 15:20:03 crc kubenswrapper[4778]: I0312 15:20:03.705097 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mnjql_14351deb-3286-4464-8eac-6bb116a9ebce/kube-rbac-proxy/0.log" Mar 12 15:20:03 crc kubenswrapper[4778]: I0312 15:20:03.819534 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mnjql_14351deb-3286-4464-8eac-6bb116a9ebce/controller/0.log" Mar 12 15:20:03 crc kubenswrapper[4778]: I0312 15:20:03.942892 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-x2n7f_2f214887-d638-42fa-aa86-1518cfae600d/frr-k8s-webhook-server/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.006523 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.178359 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.237642 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.242066 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.248212 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.379970 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.420434 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.478305 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.500904 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.715458 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-reloader/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.722287 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/controller/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.761196 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-frr-files/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.767271 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/cp-metrics/0.log" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.781890 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.857254 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5v8r\" (UniqueName: \"kubernetes.io/projected/9e32f842-16d7-484b-a241-e24ea8d3db45-kube-api-access-f5v8r\") pod \"9e32f842-16d7-484b-a241-e24ea8d3db45\" (UID: \"9e32f842-16d7-484b-a241-e24ea8d3db45\") " Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.862106 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e32f842-16d7-484b-a241-e24ea8d3db45-kube-api-access-f5v8r" (OuterVolumeSpecName: "kube-api-access-f5v8r") pod "9e32f842-16d7-484b-a241-e24ea8d3db45" (UID: "9e32f842-16d7-484b-a241-e24ea8d3db45"). InnerVolumeSpecName "kube-api-access-f5v8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:20:04 crc kubenswrapper[4778]: I0312 15:20:04.959426 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5v8r\" (UniqueName: \"kubernetes.io/projected/9e32f842-16d7-484b-a241-e24ea8d3db45-kube-api-access-f5v8r\") on node \"crc\" DevicePath \"\"" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.002925 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/kube-rbac-proxy/0.log" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.003083 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/kube-rbac-proxy-frr/0.log" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.026431 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/frr-metrics/0.log" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.238298 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/reloader/0.log" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.294924 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-54d5c4b6c7-gh4lx_a5a6d344-0a75-422d-acd9-fe8887b03110/manager/0.log" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.443620 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68f5db54d6-zstmq_6ac207b6-1710-47af-8fe9-b0c3adbce0ab/webhook-server/0.log" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.443995 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" event={"ID":"9e32f842-16d7-484b-a241-e24ea8d3db45","Type":"ContainerDied","Data":"e22dd6e038748741ff00db0c5d6a48f31b87069c92a6e6fb50d913d1489dd215"} Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.444086 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e22dd6e038748741ff00db0c5d6a48f31b87069c92a6e6fb50d913d1489dd215" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.444039 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555480-vwvb5" Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.856604 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555474-s5qjz"] Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.863964 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555474-s5qjz"] Mar 12 15:20:05 crc kubenswrapper[4778]: I0312 15:20:05.867248 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k7nvk_f2e1d11e-8f27-498d-8d45-ac0e14a796fe/kube-rbac-proxy/0.log" Mar 12 15:20:06 crc kubenswrapper[4778]: I0312 15:20:06.269147 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0be289e-351f-4101-acbd-0127a4b295dc" path="/var/lib/kubelet/pods/c0be289e-351f-4101-acbd-0127a4b295dc/volumes" Mar 12 15:20:06 crc kubenswrapper[4778]: I0312 15:20:06.544340 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k7nvk_f2e1d11e-8f27-498d-8d45-ac0e14a796fe/speaker/0.log" Mar 12 15:20:07 crc kubenswrapper[4778]: I0312 15:20:07.180496 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zxv5p_b5f035ed-2e64-4000-908f-6d0ecab1fe8d/frr/0.log" Mar 12 15:20:10 crc kubenswrapper[4778]: I0312 15:20:10.254124 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:20:10 crc kubenswrapper[4778]: E0312 15:20:10.254854 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.019133 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/util/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.178647 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/util/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.192672 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/pull/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.256433 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/pull/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.403920 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/pull/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.417533 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/util/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.420121 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wxjhd_cb93062b-8387-4eb4-8662-ecaf93146d85/extract/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.556584 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/util/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.745547 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/util/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.754090 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/pull/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.757283 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/pull/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.918686 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/pull/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.940126 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/util/0.log" Mar 12 15:20:20 crc kubenswrapper[4778]: I0312 15:20:20.963018 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rdvw6_9090029d-2f37-457b-8425-3690da177434/extract/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.082808 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-utilities/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.253842 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:20:21 crc kubenswrapper[4778]: E0312 15:20:21.254134 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.393616 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-content/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.403449 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-utilities/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.425368 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-content/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.472293 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-utilities/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.571841 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/extract-content/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.686801 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-utilities/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.909823 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-content/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.929169 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-utilities/0.log" Mar 12 15:20:21 crc kubenswrapper[4778]: I0312 15:20:21.990866 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-content/0.log" Mar 12 15:20:22 crc kubenswrapper[4778]: I0312 15:20:22.128788 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-utilities/0.log" Mar 12 15:20:22 crc kubenswrapper[4778]: I0312 15:20:22.141127 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/extract-content/0.log" Mar 12 15:20:22 crc kubenswrapper[4778]: I0312 15:20:22.395892 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hvmk8_3b062c23-5acd-430d-aa6c-24b48a725594/marketplace-operator/0.log" Mar 12 15:20:22 crc kubenswrapper[4778]: I0312 15:20:22.621507 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-utilities/0.log" Mar 12 15:20:22 crc kubenswrapper[4778]: I0312 15:20:22.838348 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-utilities/0.log" Mar 12 15:20:22 crc kubenswrapper[4778]: I0312 15:20:22.891013 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-content/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.021234 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bthl5_9098edbc-6c4b-444b-8214-5848756ec94b/registry-server/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.086728 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-content/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.120653 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fhcbf_b5b1dff9-c32b-4a91-863c-10b5ea4bc4ef/registry-server/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.217481 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-content/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.248646 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/extract-utilities/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.427620 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-utilities/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.604030 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k57lm_1d67fa18-822d-4685-a7a1-5b8b8c39c96a/registry-server/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.624564 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-content/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.640648 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-utilities/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.657353 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-content/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.844948 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-utilities/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.846469 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/extract-content/0.log" Mar 12 15:20:23 crc kubenswrapper[4778]: I0312 15:20:23.999484 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lvp8p_ca67e14c-855d-473a-99b0-fe9dabb57916/registry-server/0.log" Mar 12 15:20:26 crc kubenswrapper[4778]: I0312 15:20:26.973824 4778 scope.go:117] "RemoveContainer" containerID="deb89f96ad2640fa0674d82f73344504fdcc846f9e4815ae8eef2ce9a216dca5" Mar 12 15:20:35 crc kubenswrapper[4778]: I0312 15:20:35.254348 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:20:35 crc kubenswrapper[4778]: E0312 15:20:35.256584 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:20:46 crc kubenswrapper[4778]: I0312 15:20:46.254319 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:20:46 crc kubenswrapper[4778]: E0312 15:20:46.255009 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:20:47 crc kubenswrapper[4778]: E0312 15:20:47.195429 4778 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.32:34204->38.129.56.32:35979: write tcp 38.129.56.32:34204->38.129.56.32:35979: write: broken pipe Mar 12 15:20:47 crc kubenswrapper[4778]: E0312 15:20:47.349766 4778 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.32:34226->38.129.56.32:35979: write tcp 38.129.56.32:34226->38.129.56.32:35979: write: broken pipe Mar 12 15:21:00 crc kubenswrapper[4778]: I0312 15:21:00.255307 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:21:00 crc kubenswrapper[4778]: E0312 15:21:00.256076 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:21:11 crc kubenswrapper[4778]: I0312 15:21:11.255002 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:21:11 crc kubenswrapper[4778]: E0312 15:21:11.255957 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:21:23 crc kubenswrapper[4778]: I0312 15:21:23.253941 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:21:23 crc kubenswrapper[4778]: E0312 15:21:23.254699 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:21:35 crc kubenswrapper[4778]: I0312 15:21:35.255376 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:21:35 crc kubenswrapper[4778]: E0312 15:21:35.256171 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:21:48 crc kubenswrapper[4778]: I0312 15:21:48.257067 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:21:48 crc kubenswrapper[4778]: E0312 15:21:48.257987 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.170798 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555482-j5w56"] Mar 12 15:22:00 crc kubenswrapper[4778]: E0312 15:22:00.171700 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e32f842-16d7-484b-a241-e24ea8d3db45" containerName="oc" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.171711 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e32f842-16d7-484b-a241-e24ea8d3db45" containerName="oc" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.171894 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e32f842-16d7-484b-a241-e24ea8d3db45" containerName="oc" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.172691 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.175858 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.176503 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.177086 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.188304 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555482-j5w56"] Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.314458 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn692\" (UniqueName: \"kubernetes.io/projected/343bef0b-4527-4d4b-a357-aa48cf3cbe98-kube-api-access-fn692\") pod \"auto-csr-approver-29555482-j5w56\" (UID: \"343bef0b-4527-4d4b-a357-aa48cf3cbe98\") " pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.416810 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn692\" (UniqueName: \"kubernetes.io/projected/343bef0b-4527-4d4b-a357-aa48cf3cbe98-kube-api-access-fn692\") pod \"auto-csr-approver-29555482-j5w56\" (UID: \"343bef0b-4527-4d4b-a357-aa48cf3cbe98\") " pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.452110 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn692\" (UniqueName: \"kubernetes.io/projected/343bef0b-4527-4d4b-a357-aa48cf3cbe98-kube-api-access-fn692\") pod \"auto-csr-approver-29555482-j5w56\" (UID: \"343bef0b-4527-4d4b-a357-aa48cf3cbe98\") " pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.505104 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.966995 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555482-j5w56"] Mar 12 15:22:00 crc kubenswrapper[4778]: I0312 15:22:00.978583 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 15:22:01 crc kubenswrapper[4778]: I0312 15:22:01.660555 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555482-j5w56" event={"ID":"343bef0b-4527-4d4b-a357-aa48cf3cbe98","Type":"ContainerStarted","Data":"e0962b891e1a4f0b4147284e86bd94c320819fd05c0bc84f20fdfdf90831deb3"} Mar 12 15:22:02 crc kubenswrapper[4778]: I0312 15:22:02.674621 4778 generic.go:334] "Generic (PLEG): container finished" podID="343bef0b-4527-4d4b-a357-aa48cf3cbe98" containerID="5dbc873ca44737bada83a756ec7434fa60be6ec95b9ed80179e9560e37fb59ca" exitCode=0 Mar 12 15:22:02 crc kubenswrapper[4778]: I0312 15:22:02.674871 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555482-j5w56" event={"ID":"343bef0b-4527-4d4b-a357-aa48cf3cbe98","Type":"ContainerDied","Data":"5dbc873ca44737bada83a756ec7434fa60be6ec95b9ed80179e9560e37fb59ca"} Mar 12 15:22:03 crc kubenswrapper[4778]: I0312 15:22:03.254049 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:22:03 crc kubenswrapper[4778]: E0312 15:22:03.254604 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.127378 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.307419 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn692\" (UniqueName: \"kubernetes.io/projected/343bef0b-4527-4d4b-a357-aa48cf3cbe98-kube-api-access-fn692\") pod \"343bef0b-4527-4d4b-a357-aa48cf3cbe98\" (UID: \"343bef0b-4527-4d4b-a357-aa48cf3cbe98\") " Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.317853 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/343bef0b-4527-4d4b-a357-aa48cf3cbe98-kube-api-access-fn692" (OuterVolumeSpecName: "kube-api-access-fn692") pod "343bef0b-4527-4d4b-a357-aa48cf3cbe98" (UID: "343bef0b-4527-4d4b-a357-aa48cf3cbe98"). InnerVolumeSpecName "kube-api-access-fn692". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.412879 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn692\" (UniqueName: \"kubernetes.io/projected/343bef0b-4527-4d4b-a357-aa48cf3cbe98-kube-api-access-fn692\") on node \"crc\" DevicePath \"\"" Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.700099 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555482-j5w56" event={"ID":"343bef0b-4527-4d4b-a357-aa48cf3cbe98","Type":"ContainerDied","Data":"e0962b891e1a4f0b4147284e86bd94c320819fd05c0bc84f20fdfdf90831deb3"} Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.700151 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0962b891e1a4f0b4147284e86bd94c320819fd05c0bc84f20fdfdf90831deb3" Mar 12 15:22:04 crc kubenswrapper[4778]: I0312 15:22:04.700216 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555482-j5w56" Mar 12 15:22:05 crc kubenswrapper[4778]: I0312 15:22:05.212466 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555476-5dkdj"] Mar 12 15:22:05 crc kubenswrapper[4778]: I0312 15:22:05.223785 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555476-5dkdj"] Mar 12 15:22:06 crc kubenswrapper[4778]: I0312 15:22:06.267913 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e42af641-c33e-4b80-899f-98e5d4e78dad" path="/var/lib/kubelet/pods/e42af641-c33e-4b80-899f-98e5d4e78dad/volumes" Mar 12 15:22:14 crc kubenswrapper[4778]: I0312 15:22:14.255084 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:22:14 crc kubenswrapper[4778]: E0312 15:22:14.256003 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:22:26 crc kubenswrapper[4778]: I0312 15:22:26.253734 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:22:26 crc kubenswrapper[4778]: E0312 15:22:26.254583 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:22:27 crc kubenswrapper[4778]: I0312 15:22:27.080941 4778 scope.go:117] "RemoveContainer" containerID="9e696920a26a473d829a12f6bf276893531dc1bd498cf28bdf23c8b663c144ee" Mar 12 15:22:38 crc kubenswrapper[4778]: I0312 15:22:38.255590 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:22:38 crc kubenswrapper[4778]: E0312 15:22:38.258603 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:22:53 crc kubenswrapper[4778]: I0312 15:22:53.254146 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:22:53 crc kubenswrapper[4778]: E0312 15:22:53.255218 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:23:04 crc kubenswrapper[4778]: I0312 15:23:04.254281 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:23:04 crc kubenswrapper[4778]: E0312 15:23:04.255341 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:23:05 crc kubenswrapper[4778]: I0312 15:23:05.433116 4778 generic.go:334] "Generic (PLEG): container finished" podID="050e068e-c05a-4115-8a20-381ecb7747c6" containerID="91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346" exitCode=0 Mar 12 15:23:05 crc kubenswrapper[4778]: I0312 15:23:05.433317 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" event={"ID":"050e068e-c05a-4115-8a20-381ecb7747c6","Type":"ContainerDied","Data":"91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346"} Mar 12 15:23:05 crc kubenswrapper[4778]: I0312 15:23:05.434208 4778 scope.go:117] "RemoveContainer" containerID="91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346" Mar 12 15:23:06 crc kubenswrapper[4778]: I0312 15:23:06.299020 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t8fhv_must-gather-wpn7c_050e068e-c05a-4115-8a20-381ecb7747c6/gather/0.log" Mar 12 15:23:15 crc kubenswrapper[4778]: I0312 15:23:15.254073 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:23:15 crc kubenswrapper[4778]: E0312 15:23:15.255134 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:23:23 crc kubenswrapper[4778]: I0312 15:23:23.573387 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t8fhv/must-gather-wpn7c"] Mar 12 15:23:23 crc kubenswrapper[4778]: I0312 15:23:23.575581 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="copy" containerID="cri-o://115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3" gracePeriod=2 Mar 12 15:23:23 crc kubenswrapper[4778]: I0312 15:23:23.587644 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t8fhv/must-gather-wpn7c"] Mar 12 15:23:23 crc kubenswrapper[4778]: I0312 15:23:23.987882 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t8fhv_must-gather-wpn7c_050e068e-c05a-4115-8a20-381ecb7747c6/copy/0.log" Mar 12 15:23:23 crc kubenswrapper[4778]: I0312 15:23:23.988806 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.147951 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bskd9\" (UniqueName: \"kubernetes.io/projected/050e068e-c05a-4115-8a20-381ecb7747c6-kube-api-access-bskd9\") pod \"050e068e-c05a-4115-8a20-381ecb7747c6\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.148804 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/050e068e-c05a-4115-8a20-381ecb7747c6-must-gather-output\") pod \"050e068e-c05a-4115-8a20-381ecb7747c6\" (UID: \"050e068e-c05a-4115-8a20-381ecb7747c6\") " Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.158551 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/050e068e-c05a-4115-8a20-381ecb7747c6-kube-api-access-bskd9" (OuterVolumeSpecName: "kube-api-access-bskd9") pod "050e068e-c05a-4115-8a20-381ecb7747c6" (UID: "050e068e-c05a-4115-8a20-381ecb7747c6"). InnerVolumeSpecName "kube-api-access-bskd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.252788 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bskd9\" (UniqueName: \"kubernetes.io/projected/050e068e-c05a-4115-8a20-381ecb7747c6-kube-api-access-bskd9\") on node \"crc\" DevicePath \"\"" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.452739 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/050e068e-c05a-4115-8a20-381ecb7747c6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "050e068e-c05a-4115-8a20-381ecb7747c6" (UID: "050e068e-c05a-4115-8a20-381ecb7747c6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.457024 4778 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/050e068e-c05a-4115-8a20-381ecb7747c6-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.650638 4778 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t8fhv_must-gather-wpn7c_050e068e-c05a-4115-8a20-381ecb7747c6/copy/0.log" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.651015 4778 generic.go:334] "Generic (PLEG): container finished" podID="050e068e-c05a-4115-8a20-381ecb7747c6" containerID="115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3" exitCode=143 Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.651069 4778 scope.go:117] "RemoveContainer" containerID="115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.651251 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t8fhv/must-gather-wpn7c" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.686346 4778 scope.go:117] "RemoveContainer" containerID="91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.763086 4778 scope.go:117] "RemoveContainer" containerID="115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3" Mar 12 15:23:24 crc kubenswrapper[4778]: E0312 15:23:24.763554 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3\": container with ID starting with 115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3 not found: ID does not exist" containerID="115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.763587 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3"} err="failed to get container status \"115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3\": rpc error: code = NotFound desc = could not find container \"115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3\": container with ID starting with 115cc12adee3c5d75407f3615123df70b34b4ee2bb750778748ea73d75b1e2c3 not found: ID does not exist" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.763608 4778 scope.go:117] "RemoveContainer" containerID="91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346" Mar 12 15:23:24 crc kubenswrapper[4778]: E0312 15:23:24.763891 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346\": container with ID starting with 91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346 not found: ID does not exist" containerID="91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346" Mar 12 15:23:24 crc kubenswrapper[4778]: I0312 15:23:24.763912 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346"} err="failed to get container status \"91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346\": rpc error: code = NotFound desc = could not find container \"91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346\": container with ID starting with 91a765bb1f0c8a38e71fffab266d01f82b17250fe0665f225840c64771ac6346 not found: ID does not exist" Mar 12 15:23:26 crc kubenswrapper[4778]: I0312 15:23:26.266716 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" path="/var/lib/kubelet/pods/050e068e-c05a-4115-8a20-381ecb7747c6/volumes" Mar 12 15:23:27 crc kubenswrapper[4778]: I0312 15:23:27.192358 4778 scope.go:117] "RemoveContainer" containerID="bbbdcd9b6771dbb5cbdf7ae29f037ff7ca335ae97a393f2a8c10c715cd6d06ac" Mar 12 15:23:30 crc kubenswrapper[4778]: I0312 15:23:30.255414 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:23:30 crc kubenswrapper[4778]: E0312 15:23:30.256200 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:23:45 crc kubenswrapper[4778]: I0312 15:23:45.254311 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:23:45 crc kubenswrapper[4778]: E0312 15:23:45.255774 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.503821 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5b52b"] Mar 12 15:23:49 crc kubenswrapper[4778]: E0312 15:23:49.505301 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="gather" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.505332 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="gather" Mar 12 15:23:49 crc kubenswrapper[4778]: E0312 15:23:49.505371 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="copy" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.505384 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="copy" Mar 12 15:23:49 crc kubenswrapper[4778]: E0312 15:23:49.505412 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343bef0b-4527-4d4b-a357-aa48cf3cbe98" containerName="oc" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.505428 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="343bef0b-4527-4d4b-a357-aa48cf3cbe98" containerName="oc" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.505862 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="gather" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.505902 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="050e068e-c05a-4115-8a20-381ecb7747c6" containerName="copy" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.505935 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="343bef0b-4527-4d4b-a357-aa48cf3cbe98" containerName="oc" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.509227 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.517111 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5b52b"] Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.694814 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0302835-ac50-40ca-bc24-02064f6720c0-utilities\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.695156 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldzq\" (UniqueName: \"kubernetes.io/projected/d0302835-ac50-40ca-bc24-02064f6720c0-kube-api-access-bldzq\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.695210 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0302835-ac50-40ca-bc24-02064f6720c0-catalog-content\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.796648 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0302835-ac50-40ca-bc24-02064f6720c0-utilities\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.796793 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldzq\" (UniqueName: \"kubernetes.io/projected/d0302835-ac50-40ca-bc24-02064f6720c0-kube-api-access-bldzq\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.796826 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0302835-ac50-40ca-bc24-02064f6720c0-catalog-content\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.797104 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0302835-ac50-40ca-bc24-02064f6720c0-utilities\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.797299 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0302835-ac50-40ca-bc24-02064f6720c0-catalog-content\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.820802 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldzq\" (UniqueName: \"kubernetes.io/projected/d0302835-ac50-40ca-bc24-02064f6720c0-kube-api-access-bldzq\") pod \"community-operators-5b52b\" (UID: \"d0302835-ac50-40ca-bc24-02064f6720c0\") " pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:49 crc kubenswrapper[4778]: I0312 15:23:49.853781 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:23:50 crc kubenswrapper[4778]: I0312 15:23:50.403259 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5b52b"] Mar 12 15:23:50 crc kubenswrapper[4778]: W0312 15:23:50.405386 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0302835_ac50_40ca_bc24_02064f6720c0.slice/crio-54a89ef0cd8b4fb093d17920ff45980516ffe111e030649f871975b6031a49aa WatchSource:0}: Error finding container 54a89ef0cd8b4fb093d17920ff45980516ffe111e030649f871975b6031a49aa: Status 404 returned error can't find the container with id 54a89ef0cd8b4fb093d17920ff45980516ffe111e030649f871975b6031a49aa Mar 12 15:23:51 crc kubenswrapper[4778]: I0312 15:23:51.229994 4778 generic.go:334] "Generic (PLEG): container finished" podID="d0302835-ac50-40ca-bc24-02064f6720c0" containerID="658f3579c17d5dce563f24c175a91939dba1f7d999430cb2ca1ce286f45d2937" exitCode=0 Mar 12 15:23:51 crc kubenswrapper[4778]: I0312 15:23:51.230062 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5b52b" event={"ID":"d0302835-ac50-40ca-bc24-02064f6720c0","Type":"ContainerDied","Data":"658f3579c17d5dce563f24c175a91939dba1f7d999430cb2ca1ce286f45d2937"} Mar 12 15:23:51 crc kubenswrapper[4778]: I0312 15:23:51.230296 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5b52b" event={"ID":"d0302835-ac50-40ca-bc24-02064f6720c0","Type":"ContainerStarted","Data":"54a89ef0cd8b4fb093d17920ff45980516ffe111e030649f871975b6031a49aa"} Mar 12 15:23:59 crc kubenswrapper[4778]: I0312 15:23:59.324068 4778 generic.go:334] "Generic (PLEG): container finished" podID="d0302835-ac50-40ca-bc24-02064f6720c0" containerID="0439ee2cdec34f7a395d09cccb8281543088efaa6eab1fd3af09a3ffa44df02e" exitCode=0 Mar 12 15:23:59 crc kubenswrapper[4778]: I0312 15:23:59.324226 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5b52b" event={"ID":"d0302835-ac50-40ca-bc24-02064f6720c0","Type":"ContainerDied","Data":"0439ee2cdec34f7a395d09cccb8281543088efaa6eab1fd3af09a3ffa44df02e"} Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.174222 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555484-jwvnw"] Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.175487 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.180491 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.180527 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.180754 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.186754 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555484-jwvnw"] Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.254168 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:24:00 crc kubenswrapper[4778]: E0312 15:24:00.254447 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.304500 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/146dd143-ec48-4ae2-9989-08072e1c770f-kube-api-access-89vlq\") pod \"auto-csr-approver-29555484-jwvnw\" (UID: \"146dd143-ec48-4ae2-9989-08072e1c770f\") " pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.339990 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5b52b" event={"ID":"d0302835-ac50-40ca-bc24-02064f6720c0","Type":"ContainerStarted","Data":"e8daf4e328623f8f71fbf304d484442e97e1f45de64307cc9e97d44dc3c43211"} Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.370930 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5b52b" podStartSLOduration=2.799641571 podStartE2EDuration="11.370901783s" podCreationTimestamp="2026-03-12 15:23:49 +0000 UTC" firstStartedPulling="2026-03-12 15:23:51.232251491 +0000 UTC m=+8049.680946887" lastFinishedPulling="2026-03-12 15:23:59.803511703 +0000 UTC m=+8058.252207099" observedRunningTime="2026-03-12 15:24:00.359489118 +0000 UTC m=+8058.808184564" watchObservedRunningTime="2026-03-12 15:24:00.370901783 +0000 UTC m=+8058.819597169" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.407610 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/146dd143-ec48-4ae2-9989-08072e1c770f-kube-api-access-89vlq\") pod \"auto-csr-approver-29555484-jwvnw\" (UID: \"146dd143-ec48-4ae2-9989-08072e1c770f\") " pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.430636 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/146dd143-ec48-4ae2-9989-08072e1c770f-kube-api-access-89vlq\") pod \"auto-csr-approver-29555484-jwvnw\" (UID: \"146dd143-ec48-4ae2-9989-08072e1c770f\") " pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.497569 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:00 crc kubenswrapper[4778]: I0312 15:24:00.970171 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555484-jwvnw"] Mar 12 15:24:01 crc kubenswrapper[4778]: I0312 15:24:01.352291 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" event={"ID":"146dd143-ec48-4ae2-9989-08072e1c770f","Type":"ContainerStarted","Data":"9150841d8ba3438fe4bc2bce5a696cc25787ef9bf5844155381b66cfc43c2fa8"} Mar 12 15:24:03 crc kubenswrapper[4778]: I0312 15:24:03.375070 4778 generic.go:334] "Generic (PLEG): container finished" podID="146dd143-ec48-4ae2-9989-08072e1c770f" containerID="d3066e8a82263799c1c48ebb3267168ff31eda2585e942011782a116e2d5918f" exitCode=0 Mar 12 15:24:03 crc kubenswrapper[4778]: I0312 15:24:03.375130 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" event={"ID":"146dd143-ec48-4ae2-9989-08072e1c770f","Type":"ContainerDied","Data":"d3066e8a82263799c1c48ebb3267168ff31eda2585e942011782a116e2d5918f"} Mar 12 15:24:04 crc kubenswrapper[4778]: I0312 15:24:04.753002 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:04 crc kubenswrapper[4778]: I0312 15:24:04.911928 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/146dd143-ec48-4ae2-9989-08072e1c770f-kube-api-access-89vlq\") pod \"146dd143-ec48-4ae2-9989-08072e1c770f\" (UID: \"146dd143-ec48-4ae2-9989-08072e1c770f\") " Mar 12 15:24:04 crc kubenswrapper[4778]: I0312 15:24:04.918807 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/146dd143-ec48-4ae2-9989-08072e1c770f-kube-api-access-89vlq" (OuterVolumeSpecName: "kube-api-access-89vlq") pod "146dd143-ec48-4ae2-9989-08072e1c770f" (UID: "146dd143-ec48-4ae2-9989-08072e1c770f"). InnerVolumeSpecName "kube-api-access-89vlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:24:05 crc kubenswrapper[4778]: I0312 15:24:05.014212 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89vlq\" (UniqueName: \"kubernetes.io/projected/146dd143-ec48-4ae2-9989-08072e1c770f-kube-api-access-89vlq\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:05 crc kubenswrapper[4778]: I0312 15:24:05.402845 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" event={"ID":"146dd143-ec48-4ae2-9989-08072e1c770f","Type":"ContainerDied","Data":"9150841d8ba3438fe4bc2bce5a696cc25787ef9bf5844155381b66cfc43c2fa8"} Mar 12 15:24:05 crc kubenswrapper[4778]: I0312 15:24:05.403232 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9150841d8ba3438fe4bc2bce5a696cc25787ef9bf5844155381b66cfc43c2fa8" Mar 12 15:24:05 crc kubenswrapper[4778]: I0312 15:24:05.402912 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555484-jwvnw" Mar 12 15:24:05 crc kubenswrapper[4778]: I0312 15:24:05.836821 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555478-fcjb2"] Mar 12 15:24:05 crc kubenswrapper[4778]: I0312 15:24:05.848472 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555478-fcjb2"] Mar 12 15:24:06 crc kubenswrapper[4778]: I0312 15:24:06.274054 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341f8c65-027d-48d0-b0c2-b843867c2413" path="/var/lib/kubelet/pods/341f8c65-027d-48d0-b0c2-b843867c2413/volumes" Mar 12 15:24:09 crc kubenswrapper[4778]: I0312 15:24:09.854722 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:24:09 crc kubenswrapper[4778]: I0312 15:24:09.855142 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:24:09 crc kubenswrapper[4778]: I0312 15:24:09.909887 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:24:10 crc kubenswrapper[4778]: I0312 15:24:10.537632 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5b52b" Mar 12 15:24:10 crc kubenswrapper[4778]: I0312 15:24:10.641774 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5b52b"] Mar 12 15:24:10 crc kubenswrapper[4778]: I0312 15:24:10.721583 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 15:24:10 crc kubenswrapper[4778]: I0312 15:24:10.722439 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bthl5" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="registry-server" containerID="cri-o://db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64" gracePeriod=2 Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.189688 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bthl5" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.254588 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:24:11 crc kubenswrapper[4778]: E0312 15:24:11.254858 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.361260 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-catalog-content\") pod \"9098edbc-6c4b-444b-8214-5848756ec94b\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.361374 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2cq2\" (UniqueName: \"kubernetes.io/projected/9098edbc-6c4b-444b-8214-5848756ec94b-kube-api-access-p2cq2\") pod \"9098edbc-6c4b-444b-8214-5848756ec94b\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.361476 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-utilities\") pod \"9098edbc-6c4b-444b-8214-5848756ec94b\" (UID: \"9098edbc-6c4b-444b-8214-5848756ec94b\") " Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.365781 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-utilities" (OuterVolumeSpecName: "utilities") pod "9098edbc-6c4b-444b-8214-5848756ec94b" (UID: "9098edbc-6c4b-444b-8214-5848756ec94b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.373592 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9098edbc-6c4b-444b-8214-5848756ec94b-kube-api-access-p2cq2" (OuterVolumeSpecName: "kube-api-access-p2cq2") pod "9098edbc-6c4b-444b-8214-5848756ec94b" (UID: "9098edbc-6c4b-444b-8214-5848756ec94b"). InnerVolumeSpecName "kube-api-access-p2cq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.423794 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9098edbc-6c4b-444b-8214-5848756ec94b" (UID: "9098edbc-6c4b-444b-8214-5848756ec94b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.465679 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.465747 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2cq2\" (UniqueName: \"kubernetes.io/projected/9098edbc-6c4b-444b-8214-5848756ec94b-kube-api-access-p2cq2\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.465767 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9098edbc-6c4b-444b-8214-5848756ec94b-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.487815 4778 generic.go:334] "Generic (PLEG): container finished" podID="9098edbc-6c4b-444b-8214-5848756ec94b" containerID="db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64" exitCode=0 Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.487902 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bthl5" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.487925 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bthl5" event={"ID":"9098edbc-6c4b-444b-8214-5848756ec94b","Type":"ContainerDied","Data":"db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64"} Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.488003 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bthl5" event={"ID":"9098edbc-6c4b-444b-8214-5848756ec94b","Type":"ContainerDied","Data":"7e43af4c8ac9f109aea2498c7d43bec693ffd79761be06aa8860c32373c46a08"} Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.488031 4778 scope.go:117] "RemoveContainer" containerID="db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.518472 4778 scope.go:117] "RemoveContainer" containerID="f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.536835 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.551590 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bthl5"] Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.555587 4778 scope.go:117] "RemoveContainer" containerID="f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.591943 4778 scope.go:117] "RemoveContainer" containerID="db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64" Mar 12 15:24:11 crc kubenswrapper[4778]: E0312 15:24:11.592423 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64\": container with ID starting with db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64 not found: ID does not exist" containerID="db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.592458 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64"} err="failed to get container status \"db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64\": rpc error: code = NotFound desc = could not find container \"db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64\": container with ID starting with db9178efd3232af4d713b97808176864833cbcacd596ac79e639c4e1dcb27c64 not found: ID does not exist" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.592480 4778 scope.go:117] "RemoveContainer" containerID="f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88" Mar 12 15:24:11 crc kubenswrapper[4778]: E0312 15:24:11.592849 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88\": container with ID starting with f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88 not found: ID does not exist" containerID="f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.592896 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88"} err="failed to get container status \"f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88\": rpc error: code = NotFound desc = could not find container \"f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88\": container with ID starting with f5209881605c74797474a49d590f6fd719f3b29aca37efbdd12b057d5f338a88 not found: ID does not exist" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.592927 4778 scope.go:117] "RemoveContainer" containerID="f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682" Mar 12 15:24:11 crc kubenswrapper[4778]: E0312 15:24:11.593333 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682\": container with ID starting with f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682 not found: ID does not exist" containerID="f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682" Mar 12 15:24:11 crc kubenswrapper[4778]: I0312 15:24:11.593360 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682"} err="failed to get container status \"f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682\": rpc error: code = NotFound desc = could not find container \"f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682\": container with ID starting with f6d9f45cc4cf5a401a409f223998e4fac1829a853cedf559f6486e5de7a6a682 not found: ID does not exist" Mar 12 15:24:12 crc kubenswrapper[4778]: I0312 15:24:12.271892 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" path="/var/lib/kubelet/pods/9098edbc-6c4b-444b-8214-5848756ec94b/volumes" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.164379 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-624v7"] Mar 12 15:24:13 crc kubenswrapper[4778]: E0312 15:24:13.165155 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="extract-utilities" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.165427 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="extract-utilities" Mar 12 15:24:13 crc kubenswrapper[4778]: E0312 15:24:13.165454 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="extract-content" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.165464 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="extract-content" Mar 12 15:24:13 crc kubenswrapper[4778]: E0312 15:24:13.165479 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="registry-server" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.165501 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="registry-server" Mar 12 15:24:13 crc kubenswrapper[4778]: E0312 15:24:13.165543 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="146dd143-ec48-4ae2-9989-08072e1c770f" containerName="oc" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.165552 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="146dd143-ec48-4ae2-9989-08072e1c770f" containerName="oc" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.165770 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="9098edbc-6c4b-444b-8214-5848756ec94b" containerName="registry-server" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.165800 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="146dd143-ec48-4ae2-9989-08072e1c770f" containerName="oc" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.167668 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.196178 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-624v7"] Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.311652 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-utilities\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.311908 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-catalog-content\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.312012 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjjpp\" (UniqueName: \"kubernetes.io/projected/4ece9e25-935e-4afe-90b9-91e6e2da84b5-kube-api-access-fjjpp\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.414521 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-utilities\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.415041 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-catalog-content\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.415293 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjjpp\" (UniqueName: \"kubernetes.io/projected/4ece9e25-935e-4afe-90b9-91e6e2da84b5-kube-api-access-fjjpp\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.416851 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-utilities\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.416934 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-catalog-content\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.449065 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjjpp\" (UniqueName: \"kubernetes.io/projected/4ece9e25-935e-4afe-90b9-91e6e2da84b5-kube-api-access-fjjpp\") pod \"redhat-marketplace-624v7\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:13 crc kubenswrapper[4778]: I0312 15:24:13.512255 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:14 crc kubenswrapper[4778]: I0312 15:24:14.010370 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-624v7"] Mar 12 15:24:14 crc kubenswrapper[4778]: W0312 15:24:14.019369 4778 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ece9e25_935e_4afe_90b9_91e6e2da84b5.slice/crio-292e0499b32d3169665fd737462df8f648748b2d8088208be447da320204f9e0 WatchSource:0}: Error finding container 292e0499b32d3169665fd737462df8f648748b2d8088208be447da320204f9e0: Status 404 returned error can't find the container with id 292e0499b32d3169665fd737462df8f648748b2d8088208be447da320204f9e0 Mar 12 15:24:14 crc kubenswrapper[4778]: I0312 15:24:14.526541 4778 generic.go:334] "Generic (PLEG): container finished" podID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerID="add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f" exitCode=0 Mar 12 15:24:14 crc kubenswrapper[4778]: I0312 15:24:14.526646 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerDied","Data":"add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f"} Mar 12 15:24:14 crc kubenswrapper[4778]: I0312 15:24:14.526898 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerStarted","Data":"292e0499b32d3169665fd737462df8f648748b2d8088208be447da320204f9e0"} Mar 12 15:24:15 crc kubenswrapper[4778]: I0312 15:24:15.543064 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerStarted","Data":"98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261"} Mar 12 15:24:16 crc kubenswrapper[4778]: I0312 15:24:16.557676 4778 generic.go:334] "Generic (PLEG): container finished" podID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerID="98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261" exitCode=0 Mar 12 15:24:16 crc kubenswrapper[4778]: I0312 15:24:16.557764 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerDied","Data":"98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261"} Mar 12 15:24:17 crc kubenswrapper[4778]: I0312 15:24:17.571208 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerStarted","Data":"bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c"} Mar 12 15:24:17 crc kubenswrapper[4778]: I0312 15:24:17.594331 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-624v7" podStartSLOduration=2.113434496 podStartE2EDuration="4.594312033s" podCreationTimestamp="2026-03-12 15:24:13 +0000 UTC" firstStartedPulling="2026-03-12 15:24:14.528474712 +0000 UTC m=+8072.977170158" lastFinishedPulling="2026-03-12 15:24:17.009352269 +0000 UTC m=+8075.458047695" observedRunningTime="2026-03-12 15:24:17.588661873 +0000 UTC m=+8076.037357299" watchObservedRunningTime="2026-03-12 15:24:17.594312033 +0000 UTC m=+8076.043007429" Mar 12 15:24:23 crc kubenswrapper[4778]: I0312 15:24:23.512985 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:23 crc kubenswrapper[4778]: I0312 15:24:23.514431 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:23 crc kubenswrapper[4778]: I0312 15:24:23.564730 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:23 crc kubenswrapper[4778]: I0312 15:24:23.687617 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:23 crc kubenswrapper[4778]: I0312 15:24:23.835592 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-624v7"] Mar 12 15:24:24 crc kubenswrapper[4778]: I0312 15:24:24.260012 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:24:24 crc kubenswrapper[4778]: E0312 15:24:24.260447 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:24:25 crc kubenswrapper[4778]: I0312 15:24:25.653239 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-624v7" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="registry-server" containerID="cri-o://bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c" gracePeriod=2 Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.160340 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.305803 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjjpp\" (UniqueName: \"kubernetes.io/projected/4ece9e25-935e-4afe-90b9-91e6e2da84b5-kube-api-access-fjjpp\") pod \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.306044 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-utilities\") pod \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.306083 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-catalog-content\") pod \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\" (UID: \"4ece9e25-935e-4afe-90b9-91e6e2da84b5\") " Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.307097 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-utilities" (OuterVolumeSpecName: "utilities") pod "4ece9e25-935e-4afe-90b9-91e6e2da84b5" (UID: "4ece9e25-935e-4afe-90b9-91e6e2da84b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.323158 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ece9e25-935e-4afe-90b9-91e6e2da84b5-kube-api-access-fjjpp" (OuterVolumeSpecName: "kube-api-access-fjjpp") pod "4ece9e25-935e-4afe-90b9-91e6e2da84b5" (UID: "4ece9e25-935e-4afe-90b9-91e6e2da84b5"). InnerVolumeSpecName "kube-api-access-fjjpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.354589 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ece9e25-935e-4afe-90b9-91e6e2da84b5" (UID: "4ece9e25-935e-4afe-90b9-91e6e2da84b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.408510 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjjpp\" (UniqueName: \"kubernetes.io/projected/4ece9e25-935e-4afe-90b9-91e6e2da84b5-kube-api-access-fjjpp\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.408564 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.408576 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ece9e25-935e-4afe-90b9-91e6e2da84b5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.666101 4778 generic.go:334] "Generic (PLEG): container finished" podID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerID="bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c" exitCode=0 Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.666143 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerDied","Data":"bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c"} Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.666174 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-624v7" event={"ID":"4ece9e25-935e-4afe-90b9-91e6e2da84b5","Type":"ContainerDied","Data":"292e0499b32d3169665fd737462df8f648748b2d8088208be447da320204f9e0"} Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.666206 4778 scope.go:117] "RemoveContainer" containerID="bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.666238 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-624v7" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.700218 4778 scope.go:117] "RemoveContainer" containerID="98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.716629 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-624v7"] Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.726220 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-624v7"] Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.744689 4778 scope.go:117] "RemoveContainer" containerID="add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.774784 4778 scope.go:117] "RemoveContainer" containerID="bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c" Mar 12 15:24:26 crc kubenswrapper[4778]: E0312 15:24:26.775327 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c\": container with ID starting with bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c not found: ID does not exist" containerID="bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.775378 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c"} err="failed to get container status \"bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c\": rpc error: code = NotFound desc = could not find container \"bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c\": container with ID starting with bd78f932a13eccc34c3846d88901fc511de9ff51eed69ead88d60dde371edd5c not found: ID does not exist" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.775410 4778 scope.go:117] "RemoveContainer" containerID="98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261" Mar 12 15:24:26 crc kubenswrapper[4778]: E0312 15:24:26.775729 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261\": container with ID starting with 98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261 not found: ID does not exist" containerID="98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.775766 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261"} err="failed to get container status \"98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261\": rpc error: code = NotFound desc = could not find container \"98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261\": container with ID starting with 98452a35aa9cb5f952df88f711fde567dcc655a81afea23308d27415d8304261 not found: ID does not exist" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.775800 4778 scope.go:117] "RemoveContainer" containerID="add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f" Mar 12 15:24:26 crc kubenswrapper[4778]: E0312 15:24:26.776104 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f\": container with ID starting with add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f not found: ID does not exist" containerID="add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f" Mar 12 15:24:26 crc kubenswrapper[4778]: I0312 15:24:26.776149 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f"} err="failed to get container status \"add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f\": rpc error: code = NotFound desc = could not find container \"add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f\": container with ID starting with add7ffa810e8138e4f5a0aab11a8dee89cf7c127a5c0de47ba3820090ae76b3f not found: ID does not exist" Mar 12 15:24:27 crc kubenswrapper[4778]: I0312 15:24:27.256655 4778 scope.go:117] "RemoveContainer" containerID="27e746629157759d4e60a414cb672470c7ab54258b384fb1bc8e845de836c293" Mar 12 15:24:28 crc kubenswrapper[4778]: I0312 15:24:28.265555 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" path="/var/lib/kubelet/pods/4ece9e25-935e-4afe-90b9-91e6e2da84b5/volumes" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.626900 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5cghz"] Mar 12 15:24:36 crc kubenswrapper[4778]: E0312 15:24:36.627573 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="registry-server" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.627587 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="registry-server" Mar 12 15:24:36 crc kubenswrapper[4778]: E0312 15:24:36.627625 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="extract-content" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.627631 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="extract-content" Mar 12 15:24:36 crc kubenswrapper[4778]: E0312 15:24:36.627647 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="extract-utilities" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.627654 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="extract-utilities" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.627863 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ece9e25-935e-4afe-90b9-91e6e2da84b5" containerName="registry-server" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.629215 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.652108 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5cghz"] Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.725410 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-catalog-content\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.725680 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcvrq\" (UniqueName: \"kubernetes.io/projected/84fd7f25-a437-4377-bbfc-e01ce99102f5-kube-api-access-mcvrq\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.725758 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-utilities\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.827218 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-catalog-content\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.827420 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcvrq\" (UniqueName: \"kubernetes.io/projected/84fd7f25-a437-4377-bbfc-e01ce99102f5-kube-api-access-mcvrq\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.827468 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-utilities\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.827761 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-catalog-content\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.827828 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-utilities\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.854483 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcvrq\" (UniqueName: \"kubernetes.io/projected/84fd7f25-a437-4377-bbfc-e01ce99102f5-kube-api-access-mcvrq\") pod \"redhat-operators-5cghz\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:36 crc kubenswrapper[4778]: I0312 15:24:36.951628 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:37 crc kubenswrapper[4778]: I0312 15:24:37.254423 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:24:37 crc kubenswrapper[4778]: E0312 15:24:37.254769 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:24:37 crc kubenswrapper[4778]: I0312 15:24:37.391327 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5cghz"] Mar 12 15:24:37 crc kubenswrapper[4778]: I0312 15:24:37.795521 4778 generic.go:334] "Generic (PLEG): container finished" podID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerID="f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4" exitCode=0 Mar 12 15:24:37 crc kubenswrapper[4778]: I0312 15:24:37.795579 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerDied","Data":"f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4"} Mar 12 15:24:37 crc kubenswrapper[4778]: I0312 15:24:37.795612 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerStarted","Data":"df127f8faf420587aa69391568c947c876a1979be6d8f6373823e337e12688b1"} Mar 12 15:24:38 crc kubenswrapper[4778]: I0312 15:24:38.807643 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerStarted","Data":"17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff"} Mar 12 15:24:41 crc kubenswrapper[4778]: I0312 15:24:41.840446 4778 generic.go:334] "Generic (PLEG): container finished" podID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerID="17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff" exitCode=0 Mar 12 15:24:41 crc kubenswrapper[4778]: I0312 15:24:41.840532 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerDied","Data":"17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff"} Mar 12 15:24:42 crc kubenswrapper[4778]: I0312 15:24:42.852857 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerStarted","Data":"a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24"} Mar 12 15:24:42 crc kubenswrapper[4778]: I0312 15:24:42.876766 4778 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5cghz" podStartSLOduration=2.386085587 podStartE2EDuration="6.876744415s" podCreationTimestamp="2026-03-12 15:24:36 +0000 UTC" firstStartedPulling="2026-03-12 15:24:37.797701426 +0000 UTC m=+8096.246396832" lastFinishedPulling="2026-03-12 15:24:42.288360264 +0000 UTC m=+8100.737055660" observedRunningTime="2026-03-12 15:24:42.868882432 +0000 UTC m=+8101.317577838" watchObservedRunningTime="2026-03-12 15:24:42.876744415 +0000 UTC m=+8101.325439821" Mar 12 15:24:46 crc kubenswrapper[4778]: I0312 15:24:46.952080 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:46 crc kubenswrapper[4778]: I0312 15:24:46.952557 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:48 crc kubenswrapper[4778]: I0312 15:24:48.009931 4778 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5cghz" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="registry-server" probeResult="failure" output=< Mar 12 15:24:48 crc kubenswrapper[4778]: timeout: failed to connect service ":50051" within 1s Mar 12 15:24:48 crc kubenswrapper[4778]: > Mar 12 15:24:49 crc kubenswrapper[4778]: I0312 15:24:49.253728 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:24:49 crc kubenswrapper[4778]: E0312 15:24:49.253927 4778 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qx88_openshift-machine-config-operator(24438fc6-dab0-4a9e-8b97-2532da76d9cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" Mar 12 15:24:57 crc kubenswrapper[4778]: I0312 15:24:57.012364 4778 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:57 crc kubenswrapper[4778]: I0312 15:24:57.083467 4778 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:57 crc kubenswrapper[4778]: I0312 15:24:57.266449 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5cghz"] Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.008009 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5cghz" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="registry-server" containerID="cri-o://a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24" gracePeriod=2 Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.475007 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.484762 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcvrq\" (UniqueName: \"kubernetes.io/projected/84fd7f25-a437-4377-bbfc-e01ce99102f5-kube-api-access-mcvrq\") pod \"84fd7f25-a437-4377-bbfc-e01ce99102f5\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.484863 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-catalog-content\") pod \"84fd7f25-a437-4377-bbfc-e01ce99102f5\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.484996 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-utilities\") pod \"84fd7f25-a437-4377-bbfc-e01ce99102f5\" (UID: \"84fd7f25-a437-4377-bbfc-e01ce99102f5\") " Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.485708 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-utilities" (OuterVolumeSpecName: "utilities") pod "84fd7f25-a437-4377-bbfc-e01ce99102f5" (UID: "84fd7f25-a437-4377-bbfc-e01ce99102f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.491944 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84fd7f25-a437-4377-bbfc-e01ce99102f5-kube-api-access-mcvrq" (OuterVolumeSpecName: "kube-api-access-mcvrq") pod "84fd7f25-a437-4377-bbfc-e01ce99102f5" (UID: "84fd7f25-a437-4377-bbfc-e01ce99102f5"). InnerVolumeSpecName "kube-api-access-mcvrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.586658 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcvrq\" (UniqueName: \"kubernetes.io/projected/84fd7f25-a437-4377-bbfc-e01ce99102f5-kube-api-access-mcvrq\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.586690 4778 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.614478 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84fd7f25-a437-4377-bbfc-e01ce99102f5" (UID: "84fd7f25-a437-4377-bbfc-e01ce99102f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 15:24:59 crc kubenswrapper[4778]: I0312 15:24:59.688781 4778 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84fd7f25-a437-4377-bbfc-e01ce99102f5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.020322 4778 generic.go:334] "Generic (PLEG): container finished" podID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerID="a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24" exitCode=0 Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.020406 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerDied","Data":"a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24"} Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.020495 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cghz" event={"ID":"84fd7f25-a437-4377-bbfc-e01ce99102f5","Type":"ContainerDied","Data":"df127f8faf420587aa69391568c947c876a1979be6d8f6373823e337e12688b1"} Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.020533 4778 scope.go:117] "RemoveContainer" containerID="a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.020426 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cghz" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.052866 4778 scope.go:117] "RemoveContainer" containerID="17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.077159 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5cghz"] Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.088337 4778 scope.go:117] "RemoveContainer" containerID="f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.091910 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5cghz"] Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.148486 4778 scope.go:117] "RemoveContainer" containerID="a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24" Mar 12 15:25:00 crc kubenswrapper[4778]: E0312 15:25:00.149222 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24\": container with ID starting with a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24 not found: ID does not exist" containerID="a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.149268 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24"} err="failed to get container status \"a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24\": rpc error: code = NotFound desc = could not find container \"a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24\": container with ID starting with a22972203933d9845965e1329ce37e119f2e59e6ad314814e87c355e00869e24 not found: ID does not exist" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.149302 4778 scope.go:117] "RemoveContainer" containerID="17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff" Mar 12 15:25:00 crc kubenswrapper[4778]: E0312 15:25:00.149862 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff\": container with ID starting with 17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff not found: ID does not exist" containerID="17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.149899 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff"} err="failed to get container status \"17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff\": rpc error: code = NotFound desc = could not find container \"17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff\": container with ID starting with 17642ed29425aaa0855b90af2c32bcd15bb32608dd5f50cc19c672c2b335f6ff not found: ID does not exist" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.149920 4778 scope.go:117] "RemoveContainer" containerID="f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4" Mar 12 15:25:00 crc kubenswrapper[4778]: E0312 15:25:00.150380 4778 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4\": container with ID starting with f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4 not found: ID does not exist" containerID="f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.150400 4778 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4"} err="failed to get container status \"f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4\": rpc error: code = NotFound desc = could not find container \"f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4\": container with ID starting with f43ae3082451efc1a94dfd26db7124dd7b4e7024343af8a61f26a1798fa14bc4 not found: ID does not exist" Mar 12 15:25:00 crc kubenswrapper[4778]: I0312 15:25:00.267015 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" path="/var/lib/kubelet/pods/84fd7f25-a437-4377-bbfc-e01ce99102f5/volumes" Mar 12 15:25:01 crc kubenswrapper[4778]: I0312 15:25:01.254712 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" Mar 12 15:25:02 crc kubenswrapper[4778]: I0312 15:25:02.041612 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"2f264e55384fd404f3791d452c96ce0c6814e91ce77fcb670a6b1d8e59491c19"} Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.157700 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555486-8vkxc"] Mar 12 15:26:00 crc kubenswrapper[4778]: E0312 15:26:00.160778 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="extract-utilities" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.161096 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="extract-utilities" Mar 12 15:26:00 crc kubenswrapper[4778]: E0312 15:26:00.161230 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="extract-content" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.161350 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="extract-content" Mar 12 15:26:00 crc kubenswrapper[4778]: E0312 15:26:00.161465 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="registry-server" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.161564 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="registry-server" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.161955 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="84fd7f25-a437-4377-bbfc-e01ce99102f5" containerName="registry-server" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.163201 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.166545 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555486-8vkxc"] Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.166664 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.171574 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.172093 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.304229 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8xgd\" (UniqueName: \"kubernetes.io/projected/702a0ec3-2c1a-40be-a56d-7e905acbf6b9-kube-api-access-s8xgd\") pod \"auto-csr-approver-29555486-8vkxc\" (UID: \"702a0ec3-2c1a-40be-a56d-7e905acbf6b9\") " pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.406776 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8xgd\" (UniqueName: \"kubernetes.io/projected/702a0ec3-2c1a-40be-a56d-7e905acbf6b9-kube-api-access-s8xgd\") pod \"auto-csr-approver-29555486-8vkxc\" (UID: \"702a0ec3-2c1a-40be-a56d-7e905acbf6b9\") " pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.436084 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8xgd\" (UniqueName: \"kubernetes.io/projected/702a0ec3-2c1a-40be-a56d-7e905acbf6b9-kube-api-access-s8xgd\") pod \"auto-csr-approver-29555486-8vkxc\" (UID: \"702a0ec3-2c1a-40be-a56d-7e905acbf6b9\") " pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.492078 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:00 crc kubenswrapper[4778]: I0312 15:26:00.987219 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555486-8vkxc"] Mar 12 15:26:01 crc kubenswrapper[4778]: I0312 15:26:01.757689 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" event={"ID":"702a0ec3-2c1a-40be-a56d-7e905acbf6b9","Type":"ContainerStarted","Data":"9506c4ffad08560fc3f076fdd1a3ab68fec4f6c6c90f0edc5304acbb1150191e"} Mar 12 15:26:02 crc kubenswrapper[4778]: I0312 15:26:02.765743 4778 generic.go:334] "Generic (PLEG): container finished" podID="702a0ec3-2c1a-40be-a56d-7e905acbf6b9" containerID="8c1ffd9f3e762d8d56ae9588cf1e6a97e4b2779167ed83f673b3bea58bf7a686" exitCode=0 Mar 12 15:26:02 crc kubenswrapper[4778]: I0312 15:26:02.765895 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" event={"ID":"702a0ec3-2c1a-40be-a56d-7e905acbf6b9","Type":"ContainerDied","Data":"8c1ffd9f3e762d8d56ae9588cf1e6a97e4b2779167ed83f673b3bea58bf7a686"} Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.124454 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.191995 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8xgd\" (UniqueName: \"kubernetes.io/projected/702a0ec3-2c1a-40be-a56d-7e905acbf6b9-kube-api-access-s8xgd\") pod \"702a0ec3-2c1a-40be-a56d-7e905acbf6b9\" (UID: \"702a0ec3-2c1a-40be-a56d-7e905acbf6b9\") " Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.198723 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702a0ec3-2c1a-40be-a56d-7e905acbf6b9-kube-api-access-s8xgd" (OuterVolumeSpecName: "kube-api-access-s8xgd") pod "702a0ec3-2c1a-40be-a56d-7e905acbf6b9" (UID: "702a0ec3-2c1a-40be-a56d-7e905acbf6b9"). InnerVolumeSpecName "kube-api-access-s8xgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.295329 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8xgd\" (UniqueName: \"kubernetes.io/projected/702a0ec3-2c1a-40be-a56d-7e905acbf6b9-kube-api-access-s8xgd\") on node \"crc\" DevicePath \"\"" Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.789887 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" event={"ID":"702a0ec3-2c1a-40be-a56d-7e905acbf6b9","Type":"ContainerDied","Data":"9506c4ffad08560fc3f076fdd1a3ab68fec4f6c6c90f0edc5304acbb1150191e"} Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.789944 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9506c4ffad08560fc3f076fdd1a3ab68fec4f6c6c90f0edc5304acbb1150191e" Mar 12 15:26:04 crc kubenswrapper[4778]: I0312 15:26:04.790020 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555486-8vkxc" Mar 12 15:26:05 crc kubenswrapper[4778]: I0312 15:26:05.222126 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555480-vwvb5"] Mar 12 15:26:05 crc kubenswrapper[4778]: I0312 15:26:05.233266 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555480-vwvb5"] Mar 12 15:26:06 crc kubenswrapper[4778]: I0312 15:26:06.266067 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e32f842-16d7-484b-a241-e24ea8d3db45" path="/var/lib/kubelet/pods/9e32f842-16d7-484b-a241-e24ea8d3db45/volumes" Mar 12 15:26:27 crc kubenswrapper[4778]: I0312 15:26:27.418769 4778 scope.go:117] "RemoveContainer" containerID="7d34b9f856d96ac0b056ec93139556664e1582951a8e260f7792f49806a39777" Mar 12 15:27:28 crc kubenswrapper[4778]: I0312 15:27:28.558269 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:27:28 crc kubenswrapper[4778]: I0312 15:27:28.558863 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:27:58 crc kubenswrapper[4778]: I0312 15:27:58.558306 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:27:58 crc kubenswrapper[4778]: I0312 15:27:58.559249 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.152423 4778 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555488-4kkmd"] Mar 12 15:28:00 crc kubenswrapper[4778]: E0312 15:28:00.153144 4778 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702a0ec3-2c1a-40be-a56d-7e905acbf6b9" containerName="oc" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.153156 4778 state_mem.go:107] "Deleted CPUSet assignment" podUID="702a0ec3-2c1a-40be-a56d-7e905acbf6b9" containerName="oc" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.153385 4778 memory_manager.go:354] "RemoveStaleState removing state" podUID="702a0ec3-2c1a-40be-a56d-7e905acbf6b9" containerName="oc" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.153974 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.157559 4778 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-6c6gl" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.157608 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.160485 4778 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.170370 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555488-4kkmd"] Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.320212 4778 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9xgt\" (UniqueName: \"kubernetes.io/projected/42dbfe18-8da1-4bec-930a-e3b7bcd92c51-kube-api-access-k9xgt\") pod \"auto-csr-approver-29555488-4kkmd\" (UID: \"42dbfe18-8da1-4bec-930a-e3b7bcd92c51\") " pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.422977 4778 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9xgt\" (UniqueName: \"kubernetes.io/projected/42dbfe18-8da1-4bec-930a-e3b7bcd92c51-kube-api-access-k9xgt\") pod \"auto-csr-approver-29555488-4kkmd\" (UID: \"42dbfe18-8da1-4bec-930a-e3b7bcd92c51\") " pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.448161 4778 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9xgt\" (UniqueName: \"kubernetes.io/projected/42dbfe18-8da1-4bec-930a-e3b7bcd92c51-kube-api-access-k9xgt\") pod \"auto-csr-approver-29555488-4kkmd\" (UID: \"42dbfe18-8da1-4bec-930a-e3b7bcd92c51\") " pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.478630 4778 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.961746 4778 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555488-4kkmd"] Mar 12 15:28:00 crc kubenswrapper[4778]: I0312 15:28:00.969915 4778 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 15:28:01 crc kubenswrapper[4778]: I0312 15:28:01.957429 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" event={"ID":"42dbfe18-8da1-4bec-930a-e3b7bcd92c51","Type":"ContainerStarted","Data":"8a4da0e264af878e82189df331aca7ee06cc91e92b71212b7b24a5e11eed7748"} Mar 12 15:28:02 crc kubenswrapper[4778]: I0312 15:28:02.969157 4778 generic.go:334] "Generic (PLEG): container finished" podID="42dbfe18-8da1-4bec-930a-e3b7bcd92c51" containerID="61f54057485efda2212c8748fb1d2f18f6df1a4a883303d5a0ce51c30f43d43c" exitCode=0 Mar 12 15:28:02 crc kubenswrapper[4778]: I0312 15:28:02.969242 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" event={"ID":"42dbfe18-8da1-4bec-930a-e3b7bcd92c51","Type":"ContainerDied","Data":"61f54057485efda2212c8748fb1d2f18f6df1a4a883303d5a0ce51c30f43d43c"} Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.302434 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.405521 4778 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9xgt\" (UniqueName: \"kubernetes.io/projected/42dbfe18-8da1-4bec-930a-e3b7bcd92c51-kube-api-access-k9xgt\") pod \"42dbfe18-8da1-4bec-930a-e3b7bcd92c51\" (UID: \"42dbfe18-8da1-4bec-930a-e3b7bcd92c51\") " Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.426435 4778 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42dbfe18-8da1-4bec-930a-e3b7bcd92c51-kube-api-access-k9xgt" (OuterVolumeSpecName: "kube-api-access-k9xgt") pod "42dbfe18-8da1-4bec-930a-e3b7bcd92c51" (UID: "42dbfe18-8da1-4bec-930a-e3b7bcd92c51"). InnerVolumeSpecName "kube-api-access-k9xgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.511567 4778 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9xgt\" (UniqueName: \"kubernetes.io/projected/42dbfe18-8da1-4bec-930a-e3b7bcd92c51-kube-api-access-k9xgt\") on node \"crc\" DevicePath \"\"" Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.992682 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" event={"ID":"42dbfe18-8da1-4bec-930a-e3b7bcd92c51","Type":"ContainerDied","Data":"8a4da0e264af878e82189df331aca7ee06cc91e92b71212b7b24a5e11eed7748"} Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.993441 4778 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a4da0e264af878e82189df331aca7ee06cc91e92b71212b7b24a5e11eed7748" Mar 12 15:28:04 crc kubenswrapper[4778]: I0312 15:28:04.992969 4778 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555488-4kkmd" Mar 12 15:28:05 crc kubenswrapper[4778]: I0312 15:28:05.404221 4778 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555482-j5w56"] Mar 12 15:28:05 crc kubenswrapper[4778]: I0312 15:28:05.418898 4778 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555482-j5w56"] Mar 12 15:28:06 crc kubenswrapper[4778]: I0312 15:28:06.276251 4778 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="343bef0b-4527-4d4b-a357-aa48cf3cbe98" path="/var/lib/kubelet/pods/343bef0b-4527-4d4b-a357-aa48cf3cbe98/volumes" Mar 12 15:28:27 crc kubenswrapper[4778]: I0312 15:28:27.550468 4778 scope.go:117] "RemoveContainer" containerID="5dbc873ca44737bada83a756ec7434fa60be6ec95b9ed80179e9560e37fb59ca" Mar 12 15:28:28 crc kubenswrapper[4778]: I0312 15:28:28.558043 4778 patch_prober.go:28] interesting pod/machine-config-daemon-2qx88 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 15:28:28 crc kubenswrapper[4778]: I0312 15:28:28.558330 4778 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 15:28:28 crc kubenswrapper[4778]: I0312 15:28:28.558371 4778 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" Mar 12 15:28:28 crc kubenswrapper[4778]: I0312 15:28:28.559050 4778 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f264e55384fd404f3791d452c96ce0c6814e91ce77fcb670a6b1d8e59491c19"} pod="openshift-machine-config-operator/machine-config-daemon-2qx88" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 15:28:28 crc kubenswrapper[4778]: I0312 15:28:28.559095 4778 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" podUID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerName="machine-config-daemon" containerID="cri-o://2f264e55384fd404f3791d452c96ce0c6814e91ce77fcb670a6b1d8e59491c19" gracePeriod=600 Mar 12 15:28:29 crc kubenswrapper[4778]: I0312 15:28:29.257768 4778 generic.go:334] "Generic (PLEG): container finished" podID="24438fc6-dab0-4a9e-8b97-2532da76d9cd" containerID="2f264e55384fd404f3791d452c96ce0c6814e91ce77fcb670a6b1d8e59491c19" exitCode=0 Mar 12 15:28:29 crc kubenswrapper[4778]: I0312 15:28:29.257836 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerDied","Data":"2f264e55384fd404f3791d452c96ce0c6814e91ce77fcb670a6b1d8e59491c19"} Mar 12 15:28:29 crc kubenswrapper[4778]: I0312 15:28:29.258739 4778 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qx88" event={"ID":"24438fc6-dab0-4a9e-8b97-2532da76d9cd","Type":"ContainerStarted","Data":"887ae2a81a8eb1736aebba9abce0a10eca55df7862c12afecb5321377d470c54"} Mar 12 15:28:29 crc kubenswrapper[4778]: I0312 15:28:29.258791 4778 scope.go:117] "RemoveContainer" containerID="28185bb0bf8713237bbead875f67f2cbfd250e5d39c0866c90d3e073957181fc" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515154555647024466 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015154555650017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154535010016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154535010015454 5ustar corecore

X頥dyN(@TCi~X3CHuwH n/u`|gUĿiZ,;@); 0+SX<](_[,b UYı= d;V# c/#Jl Y!YԐKn'r f-kro$Ag/[Fz&KjUNȀVa4Q"Tkx%akEO@Uv]]C%ᬭ:89&0yǭ~pqx 8M'X$/U 5Arܙ"1ۗ"Jt ޴CZq81< Zn :zKGR*RXug{?MWHP"Jp-插5ۀ@Et@Dl%b\.м*D!b7`b=ߑՃ?m\H,>^}2wg=[ }3 :|!1ħկx7Cz)8pBe9!jqpB8;he&6 qwXUM*V Tt>ma@YA> g>e.>nzS56H汒֭F8­(%*1R %)c%H$ɻX\*Vg;dՆ!Q?X_| %>{ҭ(զtJq+wz<8xjme(#X -s\&ΔT8,2Q%`@Z[5bk ګbkfU6,52O'XzvjWqq2(٩sJ LJ"0 W!4G2,h' #𩄅j< \*%ƁFIϭ"K` 4^@7&ׂ:T@Z12j-SJxPNEޢdƯE9fTuDYpt9SzP5^k2Vy XY[[Z%'~ҕT2/>VӕHC;'+GB0W' tgM~ۛ7]g3 !D;33_|j;7+&~} [Dcۓ?3O))J ƫ\osdX*h{XܫLuuw sDZe-`;hO L;dcwD )_2Q)^pFF%N"w՜hz+[KT)EqP]&RaِuU FW'9 m!h f*5xABV|Z}-ӏ7Eܣ7>M0p?^jȃH]ޫ1/4Sp\){P0B<\~c׉VE@K'aUI1|lZkĊFN38FnC'/jv3傒mA::VAEk}t/Uv/B:v'C 9neAEvs@V8Dn!n@tqNE\eCN ȎNL˄35>R{2Je]ji8f(6,ƌ$ /ƥZ gb,!Jp49-3w[ rwyK z=G)E(4*zhB `X)F#3`#2JzΑQjO%^?s3~qPCIR\!sRO!<,\0_/.N;,Xܞa"Oϕc0O& w;Y<<ƆsY˕Rܔ`LV^b ]E^YN+=S=UE3:jYN7Rx`HF?dW>g !x0{D_/ @$i\;"IJA\tD(U>£ܔK! uW3> eXHqH 8ǚ\.e KJW9'd[M-d ,FQ.;1{nDM2T|C&0tw0ZJB@d ǒ*/B -k.?{Ƒ ={(b n8ɾl`(i)qr~4"9V{nE KwWu@S N qHTZJ9b҇H-:8qO7rO7^-F*BN:jGϐ ! @F,4q1Vre DZsSrm?-; 3h Eg4vV- Tb%~ERc.̄K*,/=[I+Ze뿿.^kHǏ¥|Z9 r+KzW)n;+*W!>6" cÜ9 zr9ÅJ)"[YV!C7 E#LQ*C, c`C%YZN,@Q[5\.o.k?["5QbgE_5iVM(.);/,Td6> 䄢GW2$Dq/i%Nm2IBwjB)Q5.?2'gK4RpN-̘NΫ)9FeĎ_>uѠASqc*6:(P$2\Ì !HMA']w\?gmώ@oD^f}4W_/rR)i9͕cdG"=:(b*w |9d "b a Stx!pMit[?tUl.u{[ݎ)_&r0si;Gqƣ49t@':QS {-r&mKq=HAZ=,8b}gM4Rtnu}?x4HG 21 a\y EL٨1Ώ5㵂Y%ڲ0-DǙ,zp|[Oq^yRM6ɐ%3@tOVQBD"&||(PJ,֋+,k4sWy/cά廪Xw ۧ 37w{wwjY`71a%UF _g>)ի.G@uXm|sFHwhl.׼r:Og?߼[\l>of7|x@ZZAOB%wxIn%J9zboc4rK3R#RЕy(^5=^4R'tetMfEp~Gy{[# ysIw@Vr$;<=1|1WHMQ+Ӂ{t~%SXoҩhc#,jK_UBUp>}R + Q'\<AWz{'@J.*)nyZzSNSz pyp[$S]1WC`YNA !XK#3ÌA(JP0eҎ 0RjD,2U-lW[*xn2BNF9K-(B%%ղI;tQrJz2͍[Qvϥh#\BrAYikF kI _ uz+ŏ/tbӅcx rZv :BBi܂ 1Thipz^=^wJ89j9ʩv8({6y桂'sΡ̞9x) B,?N9 c|<5/\`9t>Yva) 6 ̡S4-B52*MhN[M>LZJWKV:j41e!rFc̆+6jw̫e=QQ#8&"Tȋ$=sۂ_]?l)1%P5ctKgcPl'n aV`HpF! A  /N7'z'A_ESkvAc QTa%A1AP8'4:*$ԂZ)קaՃ2] S&*RJZt2F)PPNy"%F@BkFE8(5 M *z{n +p]{EbRm* V}mͪr&^PY7[_߿&_}I/A!ү~Z,gsDےfE^bun^ ՟wm\nO-BRv_ç#FdYy w*!FD׊:8˲E.TmM:A9+z{vR(9#M B!<%Xp('MeI+Sz5^EږVZ\|T -μ c/(LHvl{JUǐ,`S r|0kSW㩕Kf(f*U>Иb/ԔP t$եP2fiLҎm@asCc> 06c> ]0B3wyCx];_>'DeXnRny^󛰞Żeupj4&. Gnq^>UPxPfHj;I{{)> ep&uYqZpPfa_$9OT${j(M:[$ #8,DNESh֥HZޢw\$br}H'3_$%cΘn/YmKg/k{]:{Y; 7Vإg[pk:+)d i~ K\*M[$K>Q-JE#HA#\!\VUs7cGwԠfEsɵMldKlYi<(,<~aIPkO6,_*`F_C{\+/ViuRNeqJc_#{\H!$yeJ%0ɀ8E1!G>$rK$.aX_gpI6mMu+5w{fRWE։ǹYXyh~V6MkoZZnpY$>DQX#N6#`8WM%׫V\\{';g=boV=T׺s> A[f _n  .\ei:-[T5n{U(nN?f@ӼYX* -Zml)E SvI+-e\`$vym[E ,zM4O@ Hf*ҮFY=ؚ2p1cp y)af%Xb,L)W&1*T!w{A﷩(g\oZO -O~sjxAmt.9叁#4>uL W\]N@\^Z^u=@nA6p- JxHˍA܀dKIYS[$<5J o +|ۏk,Tg$KEKm/)^UV`5VAC8)qiƜ ʣ먹 FBOgKsӎIsפCgJGjVk=V0E{9rE(6T 0'rIXTBVs-~Y͍}1ǕPÚ 9L%HspQ$0 r$#md02⬊QvĨ1Hj*1|;id6Ek#PA"p9aLbuɠ[A1)Nt%IꢭJRvfcc4gw#Fy0b4S?4 nbv K$ (Fha@f/zm5_!`6"e[=N.*"(g52Dz˝"p \@FɉD khĒ~$E(#>c .5+cFJCEF H ZH\^R}ے~[,1~e)\\<^+%(?}òkҕj~</4EXN&ލCd]_DDhQ-GIwǓgAxjf(({zs{GRPEDPt =Bs%R02UJ g!4D8aT " RA^ mRQ r'vy)n_vZ[B'L ΰRQ@{AH1A1daJ?#uPSE t:HsG؂ J$yk@E 3Rb^APFl FiNb6v "F*5<;VCpa*{q%6h΄Pg#@v3!J%0@L(ɑor&A\g,u:T.(,ۙHlZK7'?1L'l7@;$S?ZVJF ẒFTIbaZK0ڳ#8b=w;}6kAh)W &g(" SH8Rfd-sKP% k-JCdRK=`ZY 'փ8}9"DgIJyƂ+5'8R\A6%Th@BzMOmYZWS%8i58$OA@O%yDIDo0`ŨS. 0 nI]]\\X;xRX{633cLþF1C>ha#i*r3kvii}` Bß#N\df`%h⩬MHhUƳa\GsƣP~}gZ ij7,@oM(rBh鵌ҩ">ؓU8Y,%`B lAnXժh]i@{1L$&(bhsh_ 5 @c9k+wLO psH5{HSwܣxtrpk`PWRΝgaB'g p/+8[suW}csX0>x2<4)4LS'{ݽ+H3o A#.R"Z aEΠF9r>.6QSl!gRq}K5 话?3~?<\QPJa6 y?7R, =]p󕾻M8bˠ/ԙ7_}73V]d:mٸpT "[q8tGgvR^bg!ⷭlYʁ9Jee<8Zuw4ml&Vƭ=vt֭ IE!#ݳpCx bU%7^ hx`sDN90݇y)3Gp N߇Yt|(m[ a9SpT"n V9M#m Շ3DoԄ"3.DpJmA$M@Z9 aL~M`裶RRI08x~QkWK]&݂.Us847aZa<@21V\<ㇿ*5:X˧k\3%mxf2 o!=I%hG/U>P6 ha/Oʜ !8Ap ߥ u7KU[+w _<ٴX~q9_F™O8߇؟5] <14.‰A3…hOfsQR*+5"ݭ|Ƒ篷C4" !1!UItnجޅf%';wеbr3GǾh/11-_ƛxAWT"cR:?}C~8-)jHd⩫U Kcn?.[#gL#vW)RqV߀6FrgId{& O1F)9|͵:/fHI0SَBc912H=3AEd:TarV)`i.Flێ FuZqpuDgr/?MxU3+m9Ӌ<ո;z8hNIA T=Ғ;|1'KZp4_9Rĵ{6o+Tbx<$u #L؜|+*2gP 1a,2 ]5W۫:B`~P":#'Qxy{5۫ꊫs\J]*{Nb vی@-(t7|Ir~QzvYyB)ؓO%&RRy̔1!L]ge۔gŽ& k]!`#d/I4 KNK@Bx#?m bË\⅐M |MM[x&xD*Wk cqPymXji,4*(rTGVZ4'HD,kTL *UBotiZdN4mmޫX9%34czcJV >*ps{jd(#(.TaCu+IAq#*VGrcٛ"{ R蚴|o 8#V>-vcX-ZLteRM+qؼT%wu g>:G%jkIc8g(p=t!D1@U9-#'& r흱C!P/-<kj#EǔhiB90Lf*ʪ'(/W|.eIymqLR95wdNYegj,O1P`BT~?]s]%ۓ'9& ~Ħ}&v$}T`_X2ĊU0]7DJ^[ؕQD/UT.l7#G3$Qoً7+C$ƌL1S ^qsd2'E,qNyB>E)bX1D/"E1ZD1n‡dH'+QKKd8XDDD4&zlPEQl#JբHB/yC U\OR/,,/ǑHƼb}B HX- <)1nr}#yt,2uaY"##38m_Ӷo 1 ?jHIE继QLv_iPRS~kbJ9x 3_6G3%lKƍ|b"pH@j9`X rSv[l~@ޅ= #&8B9k!^t^s/CP}߸-Vֵ c K b uLT8F` #Xyps|L)=Z|_oT&Z!h0Vh(ri], Ҥy&8tQ>]^ZϮPŮNBuq=D+A|n re uN¬hMׅ<,HVi)< ,GǣrZ8bgbJ˕­ 0"(,[x#Ia5k6fb: =Qq@8n1jhp%1(̃aEd+ NDe072J+WT ']fLb)#*b TM0,(-Q؈|X)\ªLhA(.)apLb#<#E]TSj(H[/fӂk&8n]rd۹sS5۹ogB2_}1%q="o2$h;ZuG 46GcZZ0t\o+-&NtK5 R}?8J5,C\?3NK{q0H.*PP7lQI׉ ߜqh(Bȁ\8:# 򍛳a)`\u(;`*п^)譽6!Πߠ_o_e޲')V>o|$IySM iX0g7n=[icYD84͐ H<1` ]`iˊo*UYՄ l7WJ`W$aMm*}+3îSC李@> 謬,-Q}"N;y;kBr鄦$*r,.]Rg6J(ʕomІHϢcږ4@WpVeL%gjעk,˳fC Rf<l<ոk8ׯe=NZ1*@;1w X̗sz9O58[.@T%g8ͦBp EuZDwl0`˪o'!4دIM¦P$m$5/rX_KN ZH^w- g +o,s313~fr+@jlr)u;)F(@Ђ}"nq4>PU5ϔm6_@քLqAxR†S# LJ>}%%lsRq$)aXgrgBiXI Pw eF$%4^cD1)<%=#!PR-`a QG *:"ei 肷,:Aey4Ε*IIp;`6o mVSm7߾,ݛ#aox7|<tU-:;xQu;x pOֳQI+tZq{p ՙ{peuxaF5 G*d=z8ۂHM8/ŽV DSq~'Y#TkoACcQC,c9 0g#^j[xK/ kbRA6K'_ᅧCh_lCNuR.s|]+B#t!Er(4c),/c9JO9һ01(A!T pot =QEp u),r]3| \%cTD& mx,\J`ԒUaF_Y”QLzSmUcKzӃm PvBM50(iHgJWwVNiM&?4gbb U4xEgA3&KY͓c?qz!>50,lh=Ux~u~2y)Etn$TO*n}c$&07Ɇ)0u)hC $ eT)+2DZF Q+0Kv*nmg2'H6HsJh}o|7ɨIFM2*nFEYAT@|d!'uDQh5\RFSUEmGDmgɥB4Ε8ƌCZY[nAX*J c=RKIl$ ?$XaXNsĈcXE-@5 DDG 5 FzN1V%&yXb2R׾_к8D/tH1Xiaw1Ɣ2&% "L~APaݾ`dֵWR.8`Jɍg CH*_85fNrk:(Bl$Ah> thTlYBuDZc 5D =\2 B k<9v!: FxYРeUhCL&S# ϭrBiXmsaNw)Dg?o9T +a4CZrϫ3,r9<#aL8 = CmJ8bRN4+Ū4^*,VdO1LuJ/{;ư;%F d!|@N^l*{hSP -ducIDKM%?$UhH4?Hq¸?x(aMďD )SP'I&pU2}1IJ0'%pyPQ]oWN)RDջFcJ~]O9,FpS"X 5/ s> 2@ٵ&5t`N$F xXfPE:Uh73Le#5H1q@!mo7NIÑ GoWVE"XL.&^6k\dz/˧[* C$ =g[`&p8`cG]6 m :T-HT\PRȜP>Ք"@t*^ 7{|FW'ڠc )( -XZd>()9Z,v{Mm&BTQt9ՕbEH^X#a![@𠕎ӽ<5\{9FaHt$fC1Fi$ prgms;dūU'-C T; {Rv! Nj!8lC1G,X('P||ƙF2]i.f||$X>79^yMgݻ'ˉDl-oFרL)/kOm%J[Om>HY;*xlVrwty@VpL2Hw[ 5Aev1O)e(ˏڱRJd70+Lr|:ǜVH1!O%j|:A0VFoCE|%q,J诅^1%YUW7 26XwWW~ jJWGL}らQd|swrۊ 5fxЯϲ /mbM:锓΂;% /(m6]NƦ cO+mNج9#OVgj 宔L+^Ʉ뚐jB$e%ELǯ`9oey*bY$WNr P9*JrwLJ@|Xt p) 䒲@^uo>\m| 3ݯ$/)4튵(R)b"ٶš&V2D Ë48uKl2Oac~_ fu;ib`pO˄SOr[gm%?3`[~mޠk}0cICIi/mObڟ 2*<=i#Xi{ XOkdA]Ǥ!?U"\aڴ=eiPB׫Dg<@㧏 ڥop[߼g(w9·7j"@rim]c!8hY0*<`! {0~?@(Օр Tw6mer"h RGЙިYx^"' ItrqYqiqo+{AC y!YdDFsBWȡPK䅜uJil\"'ACbsڐKdKr0Rr$YD6ñqfQ!5Fȉ)̳9e~ra4 ؁ dX qflx/%Q34ce# Y:㓯D&橮OuuL!Nw_xyQ`ڎe@z`g'g/3%2B))0ه5%zz|Jfa܅=|"%S`JVƠɓ4N1j>ᬽ4!Q`:H0(Ս*GpH^*'wds_SO0 SW(6S# IڝD~CSF'2TNDcz }n z@wdjLFA @7 Cf /V@ H?i"U^EU˷ﮁk~qr#|nqqA"**`>BG͊D^-_Ƴͼ j=/_V*VqVEoi_߼߅O{s^ޛ c:),!La ,5::L X"(jXFe$[]YsF+xǔ>oKv<:%P̣uoHI m ˣLTV X/ٙo?O!L]$0֊u?OGvU_Fd^?31oa`X޴{ʹ7Z`IwhI ΰ6As"c\c`o/5ƸsPI)>3>}H3昪:ءqDqsFLK![h68%T0o("#As-ʪT:X)|.'xVp'6tۏBw=Uvr領O3jT :?F!_\*ϧ?K*pɧ7/jhoPԴfz.i{/䮷%|+rNzìWʨvhPjy7RL2,_jiFlۈk,vúugw6*^6zx ʩA Zo9qwjWBjG KZK_xF3ukaN3@+ ?Yf>HmV%pkجfNpӽ Dbq\aU]lv#ܴ]0rWs>]VM{kQkZ -wP6/[RLzvvaҪJ%_N=}/-3tyi dyIʹ xIe:d/s[(!\yksVM=T6^9=4V|4BX i#>C{ȅC08|/ZgIgIgIgUotW!*) E16J N#:?d޴Ԡ9I}Rk1$1a{Rіv}HRS+[GI}R҈Ԙ=DvTlzY`0J90%$r<]I }(7>|vR_1upt*dEu+}J4ts? i4_tk)ς=Io }JWoaQQ|e~XȞhMU4_*~ݪ[n1gl[NU#ݚ=7,/ u/!bPtrItko$Tgn'׌nMXȞhM闿'0[.uL'DYNq5[&cS w0UB[n1gl$PtR35a!{n96=n [.uL'DI[~_ѭ sͱ)Ջxh_X-:M[S%-wݚ=7ze5C/b-Zc36nՁNy>ۯ.wkBDsl k F7Kn1glZ*l-{gք칉fԶX 1B^DEH.uL'D6әߟftkBDlJt=>ZT }]{_ބ'T[uPZ@5ЍzfkX^m8!kﰃԌlB];ϘmORhfK &fIOHΝE{JdBׂ z`;gKJP_7 4 2ڽˤP19F=A"սB19F=Am-1+Pcs̍z{9f%D&ScnĶm昕ֲ19&=!u|^YW{%9>܄'mc>λ#k)3$sPeLYkscnBXt.,#19F= ٹ3̩xMcއ@=H1s9F=s}YO`v.,}17 !1ClxKcއ3C1 +9>܄'p޽ܐ9>ܨ'mo1,(a19&=RcOs{cB嘩f19<:.90//ü 8d=˫b:( X#)}hH0XLk>4`8,.p'Q\ &t50?pR:t9XbΊxtNF0u6q{bSB'‟aje ,ր# ­ :xP4,5^nA8 V)_1=_E[/c[V ?qDC2P9@ BWsNtтLc-UuLVʄER pV7ԉy$}ڭ1|T |"`dMbt3#Cq4>U­(^R*`Km,` &B wRVc5@bRD Z1xRgHL :6qTD?VdY!Vbe.:9]ԙJ}=Y( !JX8Hv$ES'"HjgʄPCA &}dϬo (j〽@Ut@sD;nɄPC蘲jFFjHȫr@Ţ>jM\*T0(n} !ADd3r«Ԇo`Z5.̀Ԙ9`/ EJ HXa`A(LW0'2' 18 v i[W8M(QQ(=ER|`& Aΰӌ:pJբHxI4x-t֋QDJq _<BPo_x}cY𛟝ӓ *wzz7 (1Dtx;cwПki9&[<,-:*^36Akk~^N&$#J!UPWeD\.lP|4 wc{QzV6L3RDb~') SN[3^Oz4%aϒא>3JW.gD7WR+r&qfr1lϝM?уTl? ~ː}[.\6>ALx;tE g& tA&Se,L1 $<@(2!WUtݻf!['!_w 2ZToŪM/\5/ppGi\< 'HC%[ 20WO~ͺЙ7[ T2z̦EP^9ltF0;sdn)VSp4_l?'?42K[zt`Jrpqpvy~uPd21hU<M:% VZB97T!Ox&>\?\Nn>+g-FzETm3]{Ň6(:]v~?y9292b`w{irDiJCL<%?慏&K=$ ,EA N{M#iܣ: D}k"X% m?J%͌gB{bj"x0G &44֘nN kj {E 1RWPFXB⊼~x>Kb?m~wpL)u%ГƟĤokfMJg5APX$&,n+?ۇf OU:_"wUs)v:O<&&iRi.7g>z^IU8Jo/Vx"*<4#^-܀ps?byfx܇!$bYp]9 Y~;-Qi bm` /ow5|1@utGփvIyGH?+dл'YjsQ}/P ̶gBOdt 6דkDW%c.uo2by%Ɣd#%f6R̝ͅB\a8DA4ʃwAc:T`@ys\ݸ@N< /ە "Y _[8&)aY@2CQa\ 7` ngn:@+ݟ!{Vp0M9ZwerNs(f)l&%g(֌4}e[Jն$!81L 3 Hw4$Z?q_7 xV X꠰j"Yvzw<i!pA@IL?ct$/N_vRGb PΞt~%R7Ƕ(Q1s2*EKϮ'.<5J~"ÿ$"JEȋ(U\ও<2+U`{-J}L@sAcUVfxINW}s}~P6X˙5|Q bD,)̯foXP+I:%8 |r 0NKCg{Fj@IvT BX_hD oV`Jn`EojԽ .ɥzPUJ$N%TϿjTNhMN "Y.Mg,<}2Ts»s[ fUE$iU&*,}B%P@Tz x #ǚKGkKOn2ZTQբ \8WQҙK"r֫mju*E{m7v9j@N_{l Th(^Ӧ҂nWdqWI,XPIl|st;0q&QuRJG9#u07G*7D6D r*WRǖ=b=M"B-RZS܎4YҰl,@:F[Zx!C@=ơ1Up _bSaL@w5*(s`5c/tU, Yc| )MJ4Z-L`M0^%XE10e|1&g@|CĪ\VvtpӉeШn(g3yi"Ws޷oǯGj vDj֝yܾhȎ$cȯ~d!I6hi q4wutT~Fv]:I9g9fwzf<3@Qd>^ʾ/<b8#IP4 Tv0{)UۗPf^f.le"Wsu~8 FSxsD_Ht+b邆㕗قyL$,09!c%_ʤIuu#B!̛Bx,Ke;5V6`C0M/枼Cpռi:Y@6?|Ȧf8\y> ni{tsBqdp~zՐr jBk,jJ;;ZؘC[> vՏEl4&TP턱Tûk[$^wmw|OaM+n )#jVE뷫t4ҪDQS(klXuxk7y%&ף퇢r,?01+9ճ y$ }5^<mzV;\zU]j2ݾnX^RDIx`EThKz+SvLCKygKY1+_dʸ|+w:f!P@{)+YCz4Z&Y(4W a\˜U4x ykdWQ=)ST=O[AŜts"iUmm|N\q֧1놇>_8O%{WB)qfG(mor`*2(Ձ_3 LIU_G`.b"IWm\ *L1ᐒ҆C6 Z7{ԁ( WqPvaVio4-s;.9%#t$#K4p Q$,d9:hgo L6!wkDBw l=9s)X󾔦`MJ9}9C:uܪEv]jjմ9o@VЈ-D7;KC|w, h<8뵂V6l5:QzR^}d[݆RmX6jRl΢IAޙ7vk_[4_c3Y6#w(QиFWiG{%K=E|$?l&кi͚DbXI7i,Lq.H!&I6;DebmQ7ޞQ%rSL0&2K[x뜞=Èd6~.caZ}4[sbў0q&](:<  9sF` e2 f-]E sd v{a9nYCWiU҇ݜ:x +.~gg-FzhiuVd/r?;xGTFCC!ZqK TmY/\],:-?܌ L8s@ ^07 _g+ݻݙ@k&bk pkXJ%7VŲ@"'y(R g9ߪ|1W-I},JTD_?{V@2/" Oݙ3ۍYOқMmѧLh?Y2?~̦gтE pq&-{3Ȣ}?~󑻻zz&r vv!,y(lŔڭ0Kflټӆ)U0 S$k2h7OuH;ĔC J=[봠(SǖѯR%9~{Ce瓟 暢9MYOpcZގ_lBK(ؠ(n &uy~e1aX$TcBIhp\k0^p% %/Y(̨{w=mP{qb6q@֗=o՛O]2*W'd|@Zn Tu[{5=w 8׎k kr5,q_gTN^.KT!IIjP@hB1Y)W8"%0I &O *&ԷveCA0^jGUUd28 UTɫX\Urom ˙2^Eh' =eC`f^%ٵfsV4gRhoOļ!cm8<8FKq.ENV͗!!/q:.0oѐV Ph&'5 ]e: 3u /}Ii߼]o{؋<r&6v<}1 BZn+s Eeoj=V2kѯ{!R`ohiW^gە#ص^6}E^|=8fA E89{j-ZZaGлWlڹK|UGPOuZc*reKUV[=\Cy^xtp&A#^qwa~G8mٷ=gۍga Y|K hWZp-@_;ѧt20cQgucʐhUo%lۿ8 = /=LqKk)U4d*NqϹNJ&&R%zLfE2ݬرRK7[v\:[XeV!(9Otq6{}67ly/?E@,xi\g I0h "x$`83Sk֘o *DZM80n]ߺـ^!-(6X.*]/n~ J 61Lξr (0p-l2g(zx_@A VqLz\{ǗℏӃp4JhQŔbGvTex G -zt' >A#^ybS8=nOYv]ԣҖ =+_.ƎGkZ _| ,WDs5kՙtZ>Ï}qǫ cAyK~o6 @YAwV;Z0m0ԷY&钎_[o!o8mU_5kib.nt`Ԏr[Sq(^htY +-Rn;Ssv\nnc&~ֆR0Svegmm&\",ĖXn#S߲R[WH;a7ey5xIQFõm.[`N9 `̜VWhx;Sj%? g-;Yhp~hc$iO1u˭AvҌ96 9gYm@cNgVZzb5Vݯ"'n-;34hsߞk땅Z*{=^BTn~{M &?vsDI ]ߴ)z5ڌMc)<h7۸2ؗ`1dsA46)qżt򑹪}Fj pNTv+aau2pW|_c"fy7U4B}^4,*l)E)^)w,Wn7̲:Y#nj}=ꇅh~hGyrasΌyd ` rРWTd4X. Za1dkiй*XT6胛ՍshTݔl*is1Y!/B^d+j?c Wpܦ*AT*(N -o\>(= L b%{}ѫѪ8cn@^/VEqUh (.̄)zv7;KNJ:k UI,!-JHBV8FXX-˝EW;h3밖Xk XJf U-Z6WqFV'+8x~%Q0. q,~h̓.~jmXZn/^zE`g#Ϫ\\M~G7sL.W<e5M痹:erߎ֟XTD{G?(ˇCw4"ߤqu>n2S$ ӏ  ."7twB9<]$uERlP2ר`}غsdw)ql֬BBZHuڑ(,5VZp Jz푳Jx.cc))2)lڑfj<+k.d?V,\ZJJz c c_O\$C!"9#:.ȿs-eE1@VrbIÉ&QXZщ׸i|yoKP-DvYv(/04ʂ?_EKz8G?>FᱳE _`Fi\:߯-&p~gFFwg+~:a5Gׯl'R5+l4uw6"0N jodqEjŁ߽2Y#*Kl lx;$'-FLnԆVL2_ EPA3ZDG;Q m"FD+jqVJYWJƄ|֊ q¢'4W+U<݅if1\ƷNvbװ! -H0 X \ x0j~ˈIFqǒiM1c_9}Ra}ZkC1E<ӛ BnǬ`P}}3QOάش/Y4%)yO'i @.'=igN~[]' ՜ u}$%KI|B\E/-R kS=&ψ14"<ݤ%xLX8kcﺯm~x72b<;̯O<-yJ[]ryE ±VO'̓I! y"ZGpTݿuK9Gnu1H諸ݎOEy[ڭ~CvkBB^֒)O>Lݴdb":uQE/q 9vݚW.#ֶFV;xLa#DtM4Na\AXx; WknoY[IԑS+$qTo퐛ɇxQ Uro'r4* gmߛh<=vG=}eŤG7_MMD I L߽*ZxQ^Lm~Ur).էA,u~TդT>>ݖ.2 8 1,ad) ܵkR:楿ju`Ѩw7/,#Ma%tX,_cxnaI Y6ɓaSJWff&Q ? 9C9ci6޶߾?\od[b>xX+Lƙ7^dLjP&1MX7eVmt2κ-*u k.%h;^8gP0Q[3ΙE zZ xwlEU gV&5+J8r1Wv ]J77];""ַ%G&s w~cU x0ne<^hGߧ )˵,`JwVru0BiLc Q iL("_]2Qdq?B';?ΦwC3,VfwhgNoiS{,Һῌ^=d=)B 4pb.K{Ч1;gF` RL9hۭ r㐄7lgE+e1 ՒQhЗs9/g_*4-J:[BG _TP$?[q.]UEy[i lrJ[RX\tF$HlW٠I[$*&=w aA>5p̬Gk a(oi qZ_Xr veGu14fQP%`QP|WA q %/V+e{ڜՊöIhDu).J_6>TA퉺jRmԔKJ֦X[sw9s{8X "b33֪ 0u"fVB*8A,w/Q`Gcr 9dm$sDG#Jccc"2w 6AAac!D)籬ٸ1*t3BGZ>߳r sF?Ivį*5 !'riwF-qk@J e0lE*[03oP(g)x[RuѬ#B)ݰI4kLBK@2PaM˝a& N ]1Evy3P9Lk%5K.[sz| !3\N=Rq0)+꼗s%FʉDjIc p;=iv<4WU~ r:}3s% $l)5ϳ߿'~Uz`T 6݃o5Ǹuo8OS\oqV t])oF<;S~~ H&"3dMjc!@p 73~Ls1VB"6(B%ؚ}ݟӘt㎋i}k 82&5tU(1MXK[gtpnT/Nُ㛫HO3׳` p@~Y߹107tksv"/ꑊt cAl#IRB0S-l{30wSg%s@&}j>55~_,Q,}8cmdmdmdmR*rs"i `]" JGNS'Al>!kI\9@1 _r; ':xHw❝'_ziԽe(M1Jg6XWKkR<`h 9cN 02QtkB!`(ءoIܭ *aƉ@gpBh<2"9MSk`AyAwRi]S3+fVS1U(hp. 79m4ҢmvJ jM֫ǕFWr&,¸aL)Z(0Wt!`_bqX؄v -yN-E44J2GtB'&e`eRRkܗaxt KM":@z;S +h|? U_n6UM;pmD6ѿ-_f0 vwdA*>^]w~!l0A+怳;y~x'ϫt6&M,U8t|nW@ɗozD yZ i\ *sek#7EAvFdŋ}XN\ٳ ^m-,A)3ÞHVV7UXUaLI `dDe 0\T}I(ќ_iekӍ)G_%nEcX6pjgsߣsMhY~sn΍vz * NY %Vu46+O çlب{ sHO$Q;Z(n洬4:Vguegb62*DZn1zTb7뜭'tװze8|.hU[9J3n4)oddx5=4TblzZix5~tֶ ʨG19ڵx($mIPrUO^!{o9~[x8jj3^#{Cc=m Ю|@O j`ÅtFM֗n?]7H>-}?̏Lʷ}{7WG74>ѓ ҢڵP[R\?$zʸLC[+FY>7W0hGz_zgiqޞ]meϽN%Q$1̀L!Aث2c9sɴV95@Aiͣa(5&Ü0)lUL B0m"M0p`i4NucHaDNUgr"$gA !&2hQj'Aɘ}BR~)H @'³iA9d*,Zf <)1,&+, FL}zҮHeQ]ޟnG=viԊnnL1ycθMj%K[ZY_ Uq@aR 0>ӵ]T TfL3Rtbe<}F[FJ0H`+ t׽а h 2-PHVĘ11f:GdQ8aQ#m6;4^z" BY9 }r ѩ@6 N[\גF)zz7, }az,W!SfTzQx'ţ8{lULդC\ж7{荇|RŻ#rQI ,ApW~q<>;ן>.__{NOg<~`沝ЭN开 t&J_n(V`,eBTI x:2G޻)Ҋ'n֙A͟4C30 LK&Cb~@F'3o6b2 _boA`FE2a9{F^>P$7_4bO3Fm_(]wݺuVs 5JXwTo4=aW{#!lQL;$m=UO֏>DZæG޲T/\1R?j癟6g@~Dg@;6s@ 0TsJ rd[UyKz|1R,A56 e|X{Q%WinC5!nljkf~\Q5/|caA0zt֏IӫX(g8ynoVGmu bG}|~/]'o)zoGˋG;cqI?EEfwƏ_/rs"#l:>"Q\}*eꄫdTv0"g?Q=wfBg?QR}1Ջ '4^cNAݖ:KE\#;t> ^k9l-ru95rqüiw3&B*Gi` 0RU}9z LogLkfY.j& \&iK ]DlF)3R((}̓ůC6Ǯotyuբ$cH_0ziJLDfrf63*2Ǹ.6|㦩a, L?r$pVl,d>V iO&w\}8Y6#fzUT Jpꇌ\pC^*C%!c3oinynCƘrmt)P$ #`~|%}0s@n}V|  -G4ƒI' Nf\ִĨd@EV3LBB` A#l;B4bʀZ@1>!6z#mL0\O) 0pQ粖%z\XW#k࢖&ST ݭ>ùeh\ղdE(50y}zN:IG =%s t׼T#ˍ V(mg%[+ا;?4^t|~bOC+D̼#?y3G:y`~DKCkѩ<&YLbȂ >G^`RmFMdnj?h;i.98V1Fޭʚ5 !seB}V쿎x[IZ[o ^s~SW/QIѓ'EGO:ڿ-;@[RdiAL2h7hY)b Yf]'?/ޔҮptC/ wrnQ.VzV LOcǷVhDڧ  X  } 76Myq\=َ)9S us c)\&~i (pLQw0r5NȆ5lXoCL yHõlXaĆ5@Pa3:LX Uw [kUa7 h >yd*{ .my\sH$ARH׉ܦ;w0Ohm|a0Oۅ)pՓӉ=AͪjF{>%ןx2rԆ V'vgP"W?3sM1fd'h5s#"L6g:\޵/Q ͪCo]5 ԭOTlv UtkRvboC Lo6NY:o xÆNYyës GY텣m&\xtADQs)1cf"rǔQ,sVe!L\74/8nV +Kގ8nDE>$ FH).#و9HPt$WX(U1a~/W2Vȋ YeFmlE@V6{ Hۛc<$'+Qi]I \{6>]eOp~6\k(Тre˷YGsJ3f2:A/2vgنo6;#n{[\'uI׫$}gѮ E[vx ʌœq#.Oʸg#<;ŝHV9rl7$7R={9{.\.H|byy I말2NXZZ\|W>j{Uо\i6KҐLĒuJ4|/ wWjJ\+(սUh,f H`=5Ҳ^bz18BALkV@4{@+HbÙtZMp &ΓJ[..oOcF)֌KĝS`&ƎJŅy;jyc!B݌F6Z(tn:˰w)ܮuio^:"Dտ3z2ƽ>kֻf_׈ KkZ%mcDZz+ObV?xсs8\y LqQV莟_&<Ɏmk-߅Y_CXQ|к!SK~=7*RP4H8V9td]K[v:BQI}*wKOUbTyePӰԔZcGr@EuXslx[W!2GPv2@ nHH!F'kڨH.Y >ߡsus")^ Z\nrG=?V;,&N8\P? `FƳw^^4Qn յ} @P˵76F7M2N.To*V/ubr%nfZcXƝuV#rjW99i9 ,Xś8Б1vjUh^$:TKh&bO3 %A:Ha2&csxi=q :8.^]aSz5HV.dJן;>wu/(|tRSAnN;a|R0V.AdD{v~ BQ=ș,i9etTݧsuy)ŒBӮ>ADVKё1vS䴛D<ܝ JLNxTPaaIoM<`21UUc%tdgHBΖ(פQf_9WC] J̜18p]aqJǏ~YuG܉!VjM\z`aY}| 2RIlapN2ٺ->vaE bަo>el{Y.M' $`+Sq*m>ba+q!$Ԣ#cl~,~:,pO,^`9bs5qT>~OɜX8ɾsyqq,%Zc0GK邱@}w,4sEg/BKg1GWt(N=1x=3h]hN\^8\,n$,-8} 3/N{t{_L ʙQp: r{RpC邕7}o6WoZU`qa&(Ñ %8kq#c*@~pL@=])E6xECt]¤;r9: %[ݪrL/t5-3gr Jî85jHa [3cБ1=/cDLߢoX@)jJ^gN^+,OKA2hd1,ʰōƮͼ㶚Ϲ5"Yr'D %>o;,y,ǔ(Ѝ'D3ƅ. w5ŶcۆI(d dOT44b:36Oy~W `WI ʂL$kjCVyX;-cpkmaʏ\OR~ Jj(ˏT{Xs.!w/)Rz}Mu+3$W!)'\M6dV);\M2[H3W};.Vd)?@SlJߙjY_j'ǥK#iܬZ7ergtr/8 ח89"5o{}7fປ!Gy\9LD3wjBXtjFZEO|qh]Si\K_0A@}d o~zAoD nWC1Q@S:a` ̗sol!D+>PlTJOX #<'DaE]tQ05bP4Ny;3bXRqL>T?4tZԽ) O{b=J-@R)`%נ.{ f1"ꠥ9Mª2$aXqxighwM$*}]xc;WMUl߸l4 V#c<ah3zуӈ%oM<*&XO1XJdi%l- H<#6=mq3nK BQE痗'[T.~YIrP$+svr)W͕Mm-sps'y$hXׂR|;kXZ^LWV߫RW}MVo#ޱqʏɏ7RNRe°+Z]jt10h,zy.EcJp(bE'x2zvucCeWD3""+hhqdÏZ<\}Ip#,k9ΐXul[-8n_aa҂ɗTh) `{uH݀ *_auXe,N)+/W_H\uSS$ DJEs9r~?uRO_R+SϓWrgB7~C<hhnb9d*O5_~fی EtIh|I'$B$dM]=@pEg-;/?"&aKPJ36 x?LPH%w~?D V')XO뼵ex҄#OvBe#bym8#! VQSITfWH!pTKHE͉g-~c4ŎXrtrO:qY-porirT ΄}Nj%x$&Mܪۄ11h>9$aԙy@ĜeP<鋺Zw4b6"r##_R (zT1ojDDevMՅfEz`MR*Hlq<߹|k%&@S}'#o~tc$&(U;y`%<0UB8=rL}Dl$$YۤXe4mU))3W r&ZD{RͶbVo/g<͆͠qZmGd-L~?Me(_+ݯIkEJdE >.7j$Ѧvԃ7@ )fJ`&VfS' ?Xj3{nϚ;5K_̓(φy˺Z[C;~ fvNϽs[:Y.VM_,i?O|FSDoPN`$l:{!U.4B+@'44-4k2r(d3LncVQqE[u/BQڡ翾_>?9hK "zO )V(DH)RA )2AI TFP:XHbR"qE@9 5A7m|F:hM)-V/Є(ӮW3&țo̕HSIr\M81SSLWR&VAا4L*웱27|}"sVzr]^RszO6"?cʷm" ȧk`s9ƌ |uJPVYhm^EQ%F9JQ:mWC_Vܔ&`YMTTpyiƥœ+ʁ+@"@⊀U\ߏ4rFjw^?lns̬O>h}r`k' *DxVAK^10N2,#Iɵ"नp[ljƉP;ڴFԦ/,HV" Ajp Y@fQ%[8b4hY &# *FtƑā#$4F㐲wrϠ߽,V^H# -̄Y۱h;T336!.%`cwʓOP^:xτ'Be +NӞ]v7߰&{eP6V(+ZTHc{ K< <@ !pQkPx2yhBTGLV9ʣQ$򀎓IsF1':9x@@k㘦 51M_XF5[<5%j_|쓛akBâ˹,+ ٞʢK]b f6VZDcOq\jXe_᧻쫜nO7I҆^Y)]<5O.9+,sY<5K15x%^<):MIP$…( a,XRjM֞D)o"8FrB*%w'I\kb/uXObÆIld%!y$ >i-R+; ;6BC跮kQ`-V1iht`辵ҡAE1'P:V%Whexrs/SU鐤4Dm%6H@I *q̫ڬ ycbMjt6t*~k:g4"\ThJ KcIF*F$ D#6) qI_[:6H:6>#-MV/YK)W.0r]9Q'OI:CPGOrmNP#QZ!L)Q:E>JGգtq"Aj L;4Olr$fDeX_ ׵!FӆA/>87f%Y8c#ӖHqPP1W5u&P^QYX6 S:Pf<78r2-OEz[H*Ucc=xFW@$@5[ )UpN@Ss믛64m|qLmW]E]E-(e7mPT$FqZr¼R!kpP$bPaC͠HgCβt. Ԩ_G37\(NWB݅\7yF7/(UV-FTݯqt6 "5!.ǂ/?g@b 1P(C Tu(ߡI]šk"7E]IQUs)gfU5.B*X ;gWNʓ +˼ijCN,٦VaRmv.߹H*Kg6||Ͱ3`͖Y=.èrUaDK1Sl9 U>۸9WnfL7@MnS RH AaE:l8}~0#.a]$f3lopM|^_r_I{;{;G޾v"ɇ4{vg+vvv_z7}~홣{N>f{Ż^}3vG~^8,^=8 WWgaYotXם[pb.[tk7zWOixQɤV47oE3l /5CA *ͼif )-jWS7lx^ó]޾o'vʎRG«ۇg|ja`KPrleyn"x%<ߕ]_dخ~^qˣÎ-_N1}ye4nov~|rgJ|+|b0 xbWθck/+Fm"@3WWa\8:p'o?ƛKl({5?_3Bcgt[/l_Y&^wR'~P!`a"G1O,ݝq /N˝.̞?apq ;Kg0#o ٯYox*vy۝NoAy;/'xQ'0X?}3`3]Wyp ";緀ߏ;֡˟&U @ׄySQqgtPh \~Yr S;u;(A(_ T:)$O=Rrڒ!Ӗ`LHeEklo-C'r$?jZ9T|Jč Ոϝ{F;|bt]׬,)O I~ 0|X_I_ ~2dkai1?ަCIp=75޿f|%ZWvp{vb.畜0Ωd;`gF'.8|ȩʕ%xM(h8k宊8'w>%'m܍qr7N]; P_gtBoc}=uIaX#Zj%`wRp|CwH| >"{&t Sr[7_ŋ>醂aaP.\:\qrra%kDINRy@j($EWl1Ψ[}~pct>!! 1m&l NgLA'! )Bp r4igbF?~4Gw'ӏ H%Y0S'ÐD+GFYzP &E)X#,Qs%*5tz~Gیzm[eY؞+fVvs֪%+٦9n:^6HUl@PĈ>)|=, *mVP cX^xZ֢e-Sb떵hYhYm2cHR=I+!`D dIǔ"9 k5;{ɦ.>먂#uB? N]mQjaKFaK" d0K 9EHo\I%cJRַDA N tɞo@ne21bvS1Wx.:=XEpmsgq; E $女$өyA˲w"*T=Au}WI M~PG'JIOFg䓥Gט }Gk"z[szoU)ܠ;}Eǭɥ-2ܣ=2\nMHPE5UG4<i9 5>\E먣PH wyi鋖xN vK_EK_EK__E~>k]#;?;*7ORKI5AL%@Y#&0fqeN`̧5 u|X?.Pt圄];͓uEEq-YΨ id"jɽ:dorQ-@#AKM܆ sσџ6m0aa?:G~C\aGv0}^-PsXFX^ 5qg™gw3Hr׀{\H<~G v !%FWO TaQ}pp,frCh:IP׉ (ӅSqwEJ:)ȓs&>OGMq0sχe.]K+r繄E3[&_in뫀 8&53>ғǎи!flqN0Ꜹ\j9,n}4}W4'רzkş ^֎a\<2_M\&g.KwȪAi78Xk RnKJ RYPMB&-cSS\_5( ?RF"H߶4n{ t7nݠ{ҠL3`0, ZEf)- Y rܣl8 E5PDvXmWJ7/Nx22iΉ2.%ƉI66,yrL~dm զw@~F֑JwʀM2JK[^O1'[e4lh?:lPx?m~(g/<ܯ^"ģW 40s9=ٰa}k +X2}U Ǔٟ?"ڿ xRȾ$uʖt gJ&o0HqdEOXNNB -:)^vND C6K 9ǺǦR:yQz N"p-CHX4iYIN괬N괬eun[DŽB˙cj^}r;Nt'x}3QoBCcExh|KR._͇V^|ݬjVO&|}5Ԥgkw6:L+(%˼RAV2Ӣ)S)C%/2x^ ?6A%Pq+C{;kuENuhi$:<ȷU" I"-ɢAx}WC1"7'k\o-u-RbEM3i4h6ئ4cfl{{ˡLߏ1u2)&s),8_`=ـb\:[]2%5u{~{ԫAn64n8I>L|óݭ7QSqBE sŵ-۠,);kqӚU>*3 xAh9s_ÿ q1c}iq~HkcD  I,J *j-ЏFQ3X8 @ ]JBJGSguAʩ-$h9DjjiF}&ksDBEEhvPzdbOɿQP c_::v_>w%`q"p5J*ME`>~t@UR9I2wMڨ_gwk~O,銥ӺrK2Ց~֟/=593/s1*)q~mYJzxzK,Fm3& sP2fI &BU",t} -GVK'^ B˶?ɉ` W;Ba"uw0ez5Y&A [1,mĝ+&Yʬ˾Rv>kN̵9x2} npq2=g?z V'h/^K)o"x\kz|E ʍ:\mh/DܜY2~`;'ͨL{0\՛{[ A'owvN\LuG5US !/]ko[G+fKwW?+@> f2H6_0)+)EL4E]c6 ?{OWꮮ|g5#59'cnCYc?=o7L.̶_+QB0ŚPT/ff!UvmZˆJ\?,!XZҖ-_GF%wn{Xj^rp%{LP`=<X3ϭUCRVĬZKr 6Z5UDp<,oPX`-IoK0Ma`a\H3+Lٍeu$_E߆ GQtrcvȵrKSKn¬8?ꔼ_vfI`R>aa0m;p7?L#]Z:E'`[[;ItSnbvglf=NCzkx8A$Ə% "J,AoƃF{iC~ PnMڇJ?<rvvwT*AzPٺpce*ީc*.vcPBzD6~4*oPc4XG% D%bܝy8I+!?nԯ$^x1*CRо29~rY._U`rϙ'dIz5 $5qwW'Mhq2s,.{E+`Bh8edBi֐|_{vMg ?z;z`\Mn=*_}__?>dԽ9Y#Lp} Xǀh7" ۷g3/S--oտ]%V c~Kzq?O~}WwU?vw2%bgG? U?Uބ[c~Oó.ovރ]8/i9azLkalv|} cvw'hG˧}s^Ϩ8f&ەEEtVrAHZ,c H.ƶ;GS5؁ʱ3xaܾ^J#gg2f{z;7.|)?w2Aۋd9(#>'-ӻϳ?6inPf;8]tg/>_$rB^BMdl hdE&Ȋ:.(/o'_l{NNOFE4gjp;Emg?OkeKqwvt(xZo !~YK+ze.Zx-9 tA\lNd}*.'-9TD#li=9M%C@b\KsG5ϥbY*vQWkbbI^YWk7'(펪Ŧrqxqث/B7ggteUq׾|ؾ?]s"El>]{u(#؎䴭Χ_vQ]Ԍ?ER찆Vĸ!!ĸ{"4nnLnǪVݍ:.k*%X]P& 9 Id2p>Z0X9NQ/>7զ۵~z&nV~jk{יhNtss?m_K`# ZRjVvUڿZٟܟ:KS&U!Fv Wru./O-6n/#U+{tTK+n^X&&uš ;WV{~!%zY{9a.w$bK\zGWıHo#W m8ߍDK?>EVNDۧS|"qcX0=/RL{bݿ\u)$T} zux'Ubf`qV̊MtdMu-HX 0tmMjy7mfX2{.d:ZHE=i۰VfmzgCC̤g<nM3[{?mRnvg5lږ~·ڳ?kz锾wtQyv8M.n;M,d',Dy-4h)&!4mbWs=m]lWAՏ7]I ۑNOXe&*"=hR/_Cg9.mš6 ju Y&5  TJN$s%Ĭ|/mn v*j'ֵfZi UWcR4K+YbkM#]U ^/HqNR*QksnٌQA'QmJ:-#(vwqmo/tz {Ѿb׃vXT^jJ aLUBx`" J`n{H R~BON7T\e"|5wlrcR+)( ZH:QX (aL9:9#օc3Ŭea@JJ)"]Ekl₣$9TH^  e{JPv9? iQCeCPukLޟuLu^;i8g쬍˙Z{} N-:/TLEA;blJxKl%%)^VecmN;T*MGh*_@LqlI$.>=Ҩ4=txUG$fQ•q:6Fb'WFO&3+"nB]^Q3cuF(Zy%봏CX:IIKnԬfI-a& 0^WŪ))i^-bL2d`5fF lQRd$%Np5% m[$3<#,M9*k^L׸j)28XGDoN IUm&R6`[V=Z\j)qswv6^0h-{-̣H-$mhamHVnDZ) $Ȕ[R!Vc qjdFs>v 'om-M3#\OڥwW-o1h}bj7խ컠TA4&܂5sUAJ2)*GDi Na/YXnTUot2ϫ:Yu`[NiM{$P=l*! L 28.T\Zr%\9XxY.JaHf`NaC)I-tL} fav6as>j]GW?_$xuyRK] ~G,QBtAo#i^%NV;Yq$Drphg_S0ދTXc0bˊ3S ֊}Elژ6pP.D-=.m-~;GhcŭvjC p{;ƒZ>R^7 cځhDr's~a'Cha6N0P۫C%) +j$Lf]*`͠a:MG6"{XMفtȀ'i3){K"H96V;@S&_^b.`: [A(6zrt0mvhUo/]ow*fia8h"OSmʎ-0g0Ax%>[ّ~F=D[MHRA۞ pSEt1+lL qʟW==tnS-MY`'Hu/Z 77}(#~bc&`w{9ϥpjme]B ؁s:HXCv~>0W )R5AhSil|K6P^l$!J"L9HB`&[ٻƑcW$9WyX"}9F TKMIԭ&)>"_U׭[PT\pWkNlP1p5ONtt995匡TȾ0F CC:PI2J%J3U}!j_QBFJwÖ&2VJXP\)prk T9pр;# I2U)Ӟ .3D Nrߣ:.TqM4qqX4ٹvBhWQ|t0z"`CDͺD\"g*\4}(xDscTHQnhKsV֔7W7?e^r9h޴nI D7 Bh.u v7{(* NJ0}dybP!!T %;ON%B5!yMQ$DՅDRMa!0R/E5̏"CEs ohbp;&M~N@Vg R$vuy]!0ܺ{yuX Թw A>+'>eJS;eǻ#7!|CL }ϧ hnxg`<!V@w A' ]C4efA["i}O瓿ݭn|[ag, ~2sB]s{w;k9Һ?Ga$]qQ?ݻXדMQiWqpvѽf#W?;V_mO3hԟә>rHhA@p'8kZ#ևʄK]b9qU8&L>ryp'wy\KD.fqB[֏@)d+!H*_5|ya|ՃZN'z9R3dNDnͺ?W\S[:.R(L޺íG$kv[sWzv^}0)]%/]T[:IYCy-[3Ux{̗(r[;?@he kMt7ﶞNꞬ)a tQ I9U,D&+)u:iIg5F)d7YvT5Y[^}A- BjͿA,̢3hE(DtZ-@VJ|4/EMKocbSX؟iP %셴%F,/iaduX7N<{''Q+vvn^b+hO=[ߊxmw qxX"93ZF}Z|Ҷ='uœ3aK'gp1܏Dn=抂U m"j^o[{;Ed=;d 7eGo5-=!8scx1s>O9gz[gtd*TA*^RT(//E*St>e޹\O2+y:&}iWn_ ~6Uu/OR,6u +MIrmTV(iZ,e"’.'?Ctz?X)(ľo"8ieJm0s2h+/v=8._& MVcmՆTM3o/ ЎT8-^IIs ȉR!h憁LS BUHzw4 USC`h SJCd2s-&)̇2F`4Ѽ3"JR7qg1 hvW90۞YǪ<{Oޞ +<5Vc'/4% ċ5[S< @ i+ĝeTM\ OsxuJ_ Q-8e~Њ,î4mݽE P&EkJL2hC5?ޤӎ#ó]sQd^N]צI9CDQJrrQdIhUtw2 Nnʋ[cb‍R웿řw7Dy:RM;㼯QrEcwi ŽwFuo w3؟t9%@ҳV/{] ݕ] U5CW-2 d 6EJش(( ʴ&nv4IKb_UXt;2<@VKQN,~Nr{sC9iBSt}M{ZsӕNW:]9tUNU@ର6%MujK)MII2AIAY(3*et*f$)޺!EQ!R)]O&vz}@ӵ$%-l~%_pƑ+ w)^.cTtGwXѾ"] ~zEM0HՅ0:ɛ*l}wBŊ2AJ.2i I4,-'+O~U6y<6LjX!zd qŇݷtZ+^||+.rv4r/n?H9ØixDZGȞӲ1Y}2zy+f[%G NXce 0)eӲceqh٤=A]݀?qA%cD;ImE^ȝ]JR5__Z7- #k"n݂&#Dftg`t|ք3I-E=&".Le^ˆuL!.l@Voc?m&3qI+OsB¸`mgԃЅrvxQj2/ !Ҭ^M^scd̨IDS)D$nn "UnwF1fwqZtнنY 5TS-!6V9O<1D*H5`nO: /Nx; )UrS>1Qz gfNjOB p -b/(z)5 0eØ\5`FHvMħd[aS˜VL@70V_!ּ32S@aɡ$Q|1-\7+MiZ :)'q4pey*W謿o9>juI`Z[_[ Y8RK,0* :%_qjS ;:޼8vT·ЏJSCa ]:;n5t٘wj0Ii}?wӛ8eSshčMhčMy|F~> |Hxo.]3ڛag=Tt8U|Fk;3=r25BU::`eg7;yrWu^U;zk V0$Ms*MJ@kk,^*Bu?I%+HKu싢ƾ|i6@?<)s L)g/?m\rTKw<|Ho/Lrue RfC [:ƒh5+_BtM}h. vΗW-fJ8\<c蔫 mH]qc#boXCޤBqF7x<Z)N )%1ՌTZyNdܢO7<@5flfspNpn[ܧxi3b # 1(F%ƆDEƒG?;zsk6ͼ=xv8{6}\|Pp)l*): Y=7žP!!˴-QeIGNZ ?,F/C]bQu-r?}Z[?.V@Jk=m}{B1)i/=-9lstWu~'ꩣhtz8 s=CwM;xlp[tlxӣGㅼV0oXrŌ y;P B VPLcD/&4p*u4Ա,{r6q^F1_fK]B ҏ.8o2zB럤Mq'p6s&wXP! M聰y׽o8v0r vVb9\@~mNw_4GJzf$_GЕ9yfJ:.bj~l u=jsWagqw* PR",1W1_z]yBavfAS`'Oc^`Y]B˪_w+DusOB2]'bkv IcY,(\˚BTTW̺@uqk|9RTVTb1+19JWż rMr~* ZEv蓍/ 00sv׊2lȰ7Mynf$JbCviMWqR!;9_ig]!Yt@t]A{2 w<~? ?IPc)9VmVROLY[uU+5؄=4> sEESJ Gڬ46))[tnd$S5\5#Jg (H2N%É1w~0}'ڨb\d"?Xo@t{C84Oi ѝ(ir32P2RWvhRFi\hמ@}m4Pg O\P'|5f偒2Gսq,ZnӕD)[+[wـy#.=N5Sf,.fd= *{f>b ".e]zӧ<HʴI3 U. =1#K_6EZLv9H7QƵH+H52VLTtGZjzYAZm&iA{>'2 $$e+ƷفR)*O79am?8?頡WA`.H&q&FuK H!IfeCR8yӖ_?Ysƴ,m{x_| ]eRܪ UL*[--y^$25 K%r̾+%u2sR':#^z$!ZR!d+IlZ<'iq=oIGKJJ8. BX3:X?f̞<5l0{V;=N[ޤ䍭lVwI\};8e 0y8a{ʪde~!|*>&I5*#jk$d3Ɓ;WX $y]&TJ7t;j+ͮ[vcɲ~+KfoeGl aD'1@v;P$gvِjZI5H_o6% oesJٳ`>>]c]-mݽ1c3žO)Օly䆵H%oRIʪPI xQv6nXV\ĢVQׂSzo!*9Q$W f%'|0ƅ`HRoLۉZ)~r3M;] J&DR]EX1ZMjfAR ?R aJk}:v`IMWs!>fnÊi3 ~ }`@P0DSv0i9yRO(Tܴ<%8uσ ]08bxsM(K|@& 0{ifh* k&}"{ ?uI~z$9~M5 |S.O +-E甉N |j3 >Irkģ d^E1PK MH$U9¨"T# 蜨SD,9frk˝*TfIʾ;9OclLh}v'~4m;Vj3NS3Ąٓw8 ' 8` zB?dTj35T* ǡD""%AB/Ԍ ]J\|H:Ku>o.8r˖zhEև;-X G2Im L:[d*Ϥ<|W2^&L@M{}4ZUM~xځO Ly;> n0; &)N_}5uo2 a: e@_7 cYvz^wҿ:w !훋l!@"A*S30$g/odN)}r*qr?jM33Kskq^=MήYݽ}g-?Fm/Ψ733p=bv:K;ޝ[ƏoLN{?i#7wV,11oS3o8fȍ3p݄؁[g̨ɋeϓ߇Or^uǝwCŹH _NpIi{s5M%}w'{oM;[ZҺA×A+C}xo-'$FJ B­w$W]˞[@jI.[=@6{4ə%Q z7$)ΝD|a%g%Ñ{a #`{-6^y)AZZݓJZGVTٌ3vp*vaX`[14V-$ݿS"vg{W R'=c٪W)*D~9yG1<-]g1bϧlgwKkq{y<RMlX$ γ /ݑu{I +/޼ݸ{ ۥ Rtӿߎ,5qmWF Pzp8|'{Gxg|O~Wu]*x~8' J~[y9#(vg^^ra|Iwr_ov`צ=J}=-e覣#%=k7e$sLߝvYCfnbjjġ\4d󇙓.᷂h_1;F\W:]<42E_4N  85".F`}٩; |K9iş p~l'ܙc;?Wr8f},Y3޴I__z K8klll|}{qpF }>ucÅ_ -p$wpr^gۖ;DekZZ ߾{sL:<վ?^iKׯڝدc~;ر_';>if(=Շ4r'.):Z("e/s76 pCg\UǏq؅JjܰM"sjyƙ5*ކaΗIHI350,]@8~-{+Z[x:3ύ㷂Aw㫖tq肀}g'v33fNcZI`'cF2-"ƛ^1P,jq}FC08txĚ70:w> X$y-xװwm+^Ncg%WS3&53@'˷?L < !#.Tx7s{'e$qeP>\ xpCx΋4Z}^'yjgL!Q),-4^ބgo%В&]CŻ׿Yk|Kܲ8I:A,5]K\\r䫂pfR٫^n iS?\qN?Ή% My&#6#-&9ӞU7;ycsͭ /G+01kҎmR0Pñ@%83|O;Jxa7'#F3v00q5x?0#WU9]uj= 7}?Y[i UJiRUIdM3u<<}"|z Eޛū0-;]"C%J)j"|y@|# JCBD }jIe5pV]+®~a ȮR"M]|ʥ\"B8=+rf(ZQ E+z nFa|A}_@`X{Jơ1e:+ TEprI\]x,6`!QBQAl(!L@@JΆ. ''0tgCi<%=fP#C $Gp'PMȰ n aĽ mPERGGRFЩG\"a|*|ep݋ 8҇_F  xTr8RMQ(&*D嚨\Q9!y< }GL8a70Bh#v1l8ɇVYdxT5*(,.pcL7&cZpdKb;.DV:9ZȎs;n d !wXB}G`dã(Q{>)#Brwwuo.l ~JlZ8ɯ$%1{833@ Ԝ>UL[Oo=n?II$_C!܈+u y5J(4(7M"=sޞ\_oh@); P8;=s堭)'c 7AាgbhQ\]y7~~*?n5} AA!.3?'|3?3D|}WX>``0qCOK$hYGa U>M^|ZBlVvQ~@kA<؉klmЭjCodJ˿U'%>iXq<[mMm/_gX,1䟳<|;tǻ^}7qMN.';մZZ>S(T[ICx:k*l1ޣ_i)ӊ}m1YN<1-%MftӃ=ַޯ?:_y\A f߿%1ܚ<.R&g4v\M֎TB榥ch'ol4zU`* ^6ZA}tG.BB REVesevB}+_6]c [ֿZ_&M9Xb SĘXy_U5ԌIybV{;‹|%+(9|r(NoPhC56x/WFxOƤ*2!cUҘyIZN) ^rbPã쉜1:oWO 0^6\k֌ƑKq goJ>{Sٛެ 7bBk-VYM-u-ySx,޺^{1*jCm!PACF6%]ezQHF$M/Rb=rce605s;gnX6%GMS;(nV4]]ډ7#GcB[-v]銦hJ銦hVKWԀշC^l%c XvaG;'Md=gUŅHrz˼Y y_/<)*DS;Rſ_8RJ>i~l+ dNFdn~r[@UZm ^~! x##nqZ5\}Xuq;(I1>^5*lcE(5xTqk8wCG{} ,o״Xku~ J9DgEڄ$HhGi&jH~+p^Y'?=Z )#Ɖ1< 2)*Saoj!)ɕmWWo.p/;e\xvq߮p!Tr/ZN0|îl% V^~Γ.3ƙ냘l0fO w!!`i`J#ǮRr48mĀ=`|;IӸ&^ARIT҇4-kJhMѡꇚe>M+Ub2.b\*R7]ս>sˈ*r*"2JC*gpmTR·`^ K#3w'dr2`} `+Gi8c.WNqaAK '/@3!#/<('GzYV;XD+|+NJ.b`YB"ru1 'Ex6B[].Z,}ٟܡzSxlmȐq~)ɋ˩!RSC`kt5O9V,JNuwS2yYf+Y)Ɔ̔Uɰ؅VK=7%16,~|Lc["Z) ^ZCd\NAu2<УUTͣbjb\qcMKc\ u2[X[R(eVvoMYTZ2VzZISS)prhv ̸P[VҀ`ʶo@)wӬܬ D̡6OAnt47o1f-ĩfc,&Oؽ<@-ӒK;X#EdDf * 9dHA^ēmy(ܥ?4%KHO )(21uA:r(ׇ:W]LTV~ZS_$e˂/^pw&NilTB+)bm{3 |M_N.Pi;UU %!0{7P>+B%A C^Ӡ}2qek!'"F ,&۔N<2r|U.+LNxf c%]"I |^? 2+z %&Rf|J2wX1ZN_ur.Jia[2=鍣 1::m!)!(_y`9Բ$D4)ANVnԔ`] $MGQ>lQ `>'Ƙ"K#> rbo 2!L;!+g$]QV O@JA[IYH@,xڃu9DyI!( JnmYݎ +a SÛ鹰ltf5 Ӑ3 GRw83ZY5,J*(ZYt}=E!qC2Fy4AhPU1V{^D8fMPx&(Me-KWVb#1^8%aҴN?ᨉʧxɻ7`3*xl t͕m7(d,^veݯ/ۛ_C3;q*Ǽpu}7;K2x$}eW Hõ"I+Pjh7UkET{z8IK1xX,PKI3$WRa'Vp~njPq*q0w)Ay.ݲq?ď#nulh?G[-'sڕ}ceegvI L؁p^>\z;v6,P㦽 ŭ׷M}x>F*5+sͯ3GF+8'Km8bz=Dlx97=,u[BB'hg cl]GuQFRp'H`80ӡ͎g͏HΏ;?7^IQlz6J.wivXK0ц7J͇X>L 4V?vx_X;Z9|7_6+KWfG} h(S9j92 >џkq疗>#8Zm_މj^>"8Zv՜M{ȍR1ؖ{Vf`"H٩v1=*jw>zhg~ۭ<_d],}Y<9x^q.:9 ZȘ>>Dg+0C5w*?sDu6|-L pޥL(+#%%5w5jw+LY TQb}li2tG*#2tqꄋXs&)Uf=t'[E tzNVvc[h1䴘ASTIs;;rb mBh4bMUojn5?q7Q6~8 J3 ɭQ #j%]uQ15$9J)#O?J lZNPHfbdʨ %(g|t#|q<*!6$Raêx@;f_H] zGrVOpa8Rruǣɍ!@[h=&-=:O-i'5/x _ |ή:@H!\nBy.7Hp^O2 ;,^S0*Kɍ] !&c=S}N&vv7?q]F>tşЗM.εs\8ź8;:k ;XД$ ,ԑRL,MtȱL`s&eS;A%?)1`O`AIf4uF?`lR8Mqux\U]GOگ*[YW%;1<좶YbຶX{"Ǘt͋alӇ:T$YsWߵ^rYw}J0G*hnЂYfkYgֿ H35Y&`D 1VelbiɘbP樔X'm~BmڬSr+1??\FIƨT%eWŔ 3VѹB+U#2*lY5[0O:&$PtrhI{f9`bXdR I+ YP:{HAjuf+Sl9QM-hJj,LYlT&1l ա]-Cr=:,ӐvrHL"NRfIs.aS ,i'U"4ge"I*n%ћ5>!A jA A (SJՃޚ^@q=J2l]-c$(*:1%ЩIW&^/2@+B'ɀʂONHZSU)W9\eGQRc`lsC pt3 Td`,CMb^%'rg9=15\|Y/!vZ#"݉pW0BaF&xDIcPеUvݬMeVMǂk-gUhj$t@+"PHѣF9<'+fi'0N=DSux XH5%^5>/&U1n(b $9)S]iHp(5O#ϐn`t3swD(?t0 }IR읧ǂ3ToʷTݡȆdbp=v- \p Gkܳ #S^曽O&WgE^̚nL< Ua5sz;x*@}УRA8P #0Նy-܊+!!1 ")8C"eb%T[8J9&ʊ(Tf$3"I$qkc|6虼KR`};}ccDoƷ}l?eo#4KX%˄C81Τ XfpQnj>~M^Be "JX(RXl=AAkj"TuBZgOM$˔ˆ'“RRLSN3lEGm >T}/tM/ЅX!zC"5ifYQR2Re3BiCp*SKR1Ύ:S@O<2h45[J W)˸RKsJ&S`cAz Ptr քQEi&` >H0 \J`u4j4%@**$X{0 NbP# hgM,Keĸ5NZo`/2@jS|r) cMds/pܸ 1;s{KJ%S/-#T:*{LE~2ҁI7G0u q9^gTQopgg1V1=sꆏWc+5-p= Y ̗(9'So7:S$\WTDI7 [|CAw>ܣn|%FUyV0*'|&Q/(7xF.Mޛ;\P!kb;%_pI_M&H].^q߾Tl JǴ*a5UGU<=U#U!-c>Ghz܏"~1pͫhwWg]8@u#1H2)auY;Jut2\{l֪J4,46vݸ8>ɹM(̹qb JZwF"JI`0dAzSfa=$K+Q%ѕ ~ YCFqaѱi C,` Vc4tIVꝏB[IȔ#'~N*&mhaNʭ<'_)GpSkD̖6bet>p#]Z68\9:HFD"aD[=aTSdP#cͯRp&S% Ev& 㨽y=tQv&\ dwW˹n{wHЖ6$E(buE%wvv1b>pFGU7pwdB3g=Z5tvUaQMD^D!V5❑ڼ(Nnh0Eٞ3B6t<1CCрI]4 x8 3 @Vұ({ v"/D&.p/Fڲs*a,A: 4Dw:[J^9ڝG+JIgOi?+ί 5 -Y#X2YsnĄ~(t;){0Tt GA֋L 2Ȳ+Ks*k9,_}^^Ԛ,5,Ppvjj^.YQQZaڔXj F Q0THm+)K@,LmxH.(ZPGPzHji(5ǘh"8ј덆܏$mm3w*+Xf|_ Xl%҄׆t;0^PK]-rDm77BY@ `.}=< v>Lh{gfxP<ΣՖ>nz[U4=S+$gaiM{oelƹ[,rB }͜쭹cZ~*%'Ӣʁ_B|I:*xt4Ο~W,h1~5g7Vxu%,= #˰[ǫOK/fM_d^.7tUB t27MMJ˘j8V4c$YIqSoicnR?fin]2J"`U]xg'cW41n*iN7`Gu1qU؋ Qh=u;++h1Z+][ %;XcyL.\d͋|.c)|.YrҜd9cRsFeKv>1?n7y"ϒkŹѨ7x/whՍ ȍFW.gtoW=]E?>|uc>1/ظd8ʁ{wW74\Ͽ5XH{4~xﱉfMGq܃FYb_~t8~=ci?:w7 1 oKn5Ha[_`Bĉ#0TO~}ݻ!Xh&_bxc0 Xvk[7|;v{W ~t>'WBwa}T]W|&Z).aJ @DIoPW7׽_@h}H\<8w"|tzF5y$ӳgwOF w r(lkl;|_D/cV3ηM\.J)(&_c_}}bꢣZ;L[٧ᗾ=L~̹/'nŁOTD51ŭx9 69XFQ)..yAYJgpLr‡ÂpX,|8 }퍺KOP19?|87ּƔ a5o}L$nZwXsw ԭBMĒZxLn>ƐjkID>}sIcerM3fȌ.#L` ْ>&)ߨP5|T'ǖ5QCg 1팑H{vS^ :y=Dc$Ya.s3ƫy3T cw՝FP&3f,}KKeumnƿp͙t3sw v !g⍵Tas9HOP:1 E9Aۣf|2Ec7hRT`<EflOͮc,U3~uM:~]*@i,^>6:qՎl~Oڵ։5wDcż~.\D $juI?`1UPeV!kXZEj^}nu)5D Xf݊^]u*ķ]\UV`ޅσp22# ewP;iHէs50^-JQ_N?j^D2 >0:0ac?n=PK,qQ=zx)ߪ-->*9jaWJvN-Q9U f?(& LV3_g\-p)FQp9qX"[x`Ĕj8LI.˟}+&9YL< M, q%}L6niB' O6%ӘE$,(632钧K8،lCif+ y }gP+$1{$>}hI$?2uS )Aܙ4E%m$z{(jRv@sJWhJq>Z&9_U],G3B!ƹޙ206ELjPzGY`崺Ϸ_sCޓB+ɕUgl8<,}zsۋ!!UeFIa'bRPãx8ZOoZ `TFv ="kBat咹Ou0ժmʁj#%B1"MZ69D@aw(Y~m,e3ˬ4'VC=AœNa'ݼ2bn?%}#6KOI%ڋ]e0 ̟ki9S9Ò NUJB.S{S ,p.YI[LBZ *l[JjH01wfo% *X坨G XE{衔dBmy*!mj{4RhwJ+>J u4 IjNRTN7Z>FDc08z0 :a)è ɝ:Xs(/è0na"Jܭd0PR !DNfr46$`(DLqI5HL iǔD D*Zgw+|xP}zx9)JtxPi{fU0K6`"9̚˹:$ò0i3oK~\oLUq냉XGv'L㸖"M$/_Lp&5#EIxK?A/уj?Фh w@%S#1L^7=RXRk)c$Sq\ >)EL ĮE' N裌{ToujЖn@`XJ:ſ; 9k;CT/40J>\QC m>đ-*%_ɳ/5 ,!-%:W(V92Uu9=k4)Wg/Zq®q4 Kaj oIJkGHe$5\ )e7VCd,P´۫.jH.P:/c"E`\whp=_7NPCHtr W*102b člPOQ[ԟ8@O/3r'\O㡧xi<zZ=:hArjBx0e\R gI-!9礴Fxʺ3^F{ djݎ:'2@n|, Q*5A{vsBzN9D ʊeA~ַșXI,i901 S3gJAHɤchR!r$&Vd9Q-9Ldzb4(q`2D4,xL< Vi*4נ%B4h#KM8mn5יOM&Z"FJv[ôUZfLI)L<$ d0EiPV0%C,',t4ji]0)bB.cT9X: a2S3kn\Ӵ%RK s9ʔ$fd,]d FC qwň`\冊\+Tn@̀d=T ~Tj h &0j9 65؜'Yu$FjRT24ŇY̥I\]g j{Dp7veˢ U]uZ.n ye..,!lR\XM[fkmpu} !Ŗ%'Mࣄr1IZa|Dž9-wXD#y0qs xfpy-~GF7< 0_Qksg~7 j?=?x8x F敳LE,f:iW卩a{ ysrjV{k =A`|6)fmx@kvoE07,X`] U%nNFjoL?39˚xgIFZOt+b `NbD3x?2̉wjT"!Lw s={?E4X2k qb=XdžPCMh޲߉[˫ţtDgZ8\~k֟O2%^lh!f8[]>m(40/0޿@yd)AkanTrs. ]z3}?6D`=1hm-*tl~}f0qrߨՍ?68&`icqd77 'kwVӫ7FfL;O s/nMz/f8^oxޑmK Z' Q^lx%0Ȩ# eh+x8Ə%h~ \w9~i(;X$y,F2 !<͕DJZ82wYQG6$:8E$!ʑjyZ9GvkZǮ\^S0> e2dV K ڒ(g^W[W6 媙UUvQZ#ǎte\y$ݲ$ƱuK\1cFʦD \GWdnr'4H3{z`ccC J}m结X'g'3\:ITtQ` {r M@ Bwg=MF'4pJd::ҁa'B)2p*oY.rv.  QʦŒHGG%2|$9J+zI0σ,~YϹ]E^vqƮ'{ycdV~]Kцm/In&Oج E}l$%q~arYII BSd=-EݕI 99V |[*h-W 5_zm˿=G#-ԓX i48G̎Fby)P/>P 31lT$cD "X;۝Ch>.۹inUB#1BYW^D5I-~IY BZr? ܑ9kJ) Fg {:WH6v  e.ا;Kb4R@xeg㕝Wv6^gc9hbC90H-1}#_Ra(E0s*9TfLNR* H\-X{IZ5!]~nͷR̛+5b+9ț=wa+JUptBw^$bk,oU '{e9 ,ECQ4DX,Z[fe(*Kx_5mj*HS VtNէoOFa#U_`͉du[,N;+"Vr XuB`QҸ*+'}W~H.G\5[lF۫OZsOnɨ>i!r|}5Jכ?Ĺ,VIĄmll}Z}3UZZJٴl>ny=>9ܢhO{Jn5ms;hh0P`ۓlihӯ(U M{0mkھWQLD ӊ ǜs|aqI*_r Q}%@sL|_Fa@P$JElɨ)XĶ gK%,#@Ҫ@o8?z㍋GWep4wGȗΡwq`q H81)J K*)@ ?2*`>.E x`v`1֪@VXR H4.U 7$fe[;'q;©e8G>9[ƁQhT$d6߮<@!ƷKwq<[v;3kr;Ο;../_]{,cc.BYa{bUC)0׽' m# i)v\jB  ,0ġ}QGu -[r#~hassooas32S}>- 򗻳Pr08i֙ pTJ p~VKAwC1pE;sP 1F_N}F!8a*HjN,p4DMAj@-d)=?LIQ&Ϳ#H`!5@4H)D!5F#* ;0~SQg_-n`cQ;thQWFsiU(;YyB1H9Nbo DSA(&95eٵNi,Yn+ C^l!VD+*]N*o;Y-YQkiG'ū䒬Zcs3ub% Tcũnlew f;=Kҟ>n#[GKR_ZP0&Z,lv⏊;Pq<f "DlKyd Dz^2Ge;4 AV̺ghrm,ŃOg!qi築=m n gW`i5_[3c9Q.L󎯇e5c)|ufͻW@TEG"gh3ra^P /BY^<yl)c[?pyzTaH;C֝Ѫ$$ӣ>=l'13?ZJb6vvٱ"E}"0Ut1ʐT(i">_iyD6HpB*bV>Z  BJCBb Qd?bѴ؞Q'ƍVo#n:^:)q;NiK}(Msu2TS%j̃Tas;IӓnOh7 GfFoXYfE7P[ 1ۥF#!y޷i´bLf!^4AwPӐOh~ gwz>{wvF叒o4*Jfٻ_'? sgדݒp<5şg$CՋWy훷O듃V~g`<&}l٫+˷٫׿]7[}|>52Ӌ._Ùp0&:~loxϓ=Kmx;͢uf\ҧ֤x5}JH˂3#ϲ?}~x6HZdgdٷMF^%g2gn>8ZҐ}l;al{}Qgēuj%y3Nx~&ZnoxK}aZ;̓KS|0pifDhJ|7V觉!UR5'a0[o6?~0 C#m?}w܌zzLk\o?Zzttg[郟&d>6a$;;?s haIW&EBp'8|I>5"K}O]﹎|C9}au:.~|:=LfO /?u+#oeC"cJK1)#-.tO/bdt޾ܨi~jÿ=t\y,ŘKdՓks5\-u@r0zRqjlP;5^LyNR㤴Ilc2YB]o9W^{HH!ۗ,~&$Nrff%n˒խ#qVWbUu}ZcW5v1,*wc߼X߫9CDNjwz3#z3.\tR֨ۘ4g591Y;(gxT{sD2$JyDe :.4hŻɇ6G-eA+v`Z=3eYUjbVcvSr&9ʚNCd-(e2)mɕ J އ}9ޒ[|S?H.@iBT%򩘐o@aCv%e JPYA^#wqr˭ϵ#0XW^H|+ #@3?A*'UBN C8 L1xlkHݬOlԥuYnDvk];=aQ((-f3 ؗ%-Sb@{m1@̻ʈ%!>}} |!a޹ x]J2R. o޼%Eձ=uYF'CWjUUF5?ܾT/A^NuFU-x_^~)P['rX^Je/' څr,_P7oS|=EfĢM?XB_oV⿖L$ U>}?JEȾ 裟yrw:Y|]rզ ̿~+hXq [F] kKu{xSC΂eU2n:7ْ4i,_ "_kpiȰh71[Rd-Aw8niMm%D==4^k &2 TeX%K6BiJ"-HͨH#F졬="J=&0&b_iwMmM$ș^"[rlY-Y} `KPm0qtߠc?At|ZьsA+8̑T1IM";-F.6y `' ֦./HYſ$|e8Y"/@H~ `N 0EaYhЈ`"JsScćI' (!dIt9 N 6>`l]cE(Yka5*ԈdX_l~Ȟ$ 0@jwxz0I`skPeӭr ?psj =KXt⨅l&U}3`!SCuT@n@No(^1RwϽkTO:2> 'yv{QxvQx>۵ǎ'&imeنKν 42=/#۳#wqh,;"nS.Wee/JPںCrl7q~͏4׷a :aԾzDSzC=lNۡG=&qlX9]}1!=:'&x}3h rRJv1s8FC쏋]>oۆ"8pm&Hů\MA;x} 1y8}R ߓbi 2izR SavG?'GIjECOjb~e+(ཱིx`$QfѢZ8➀)S@ELBmde 7F S1V+90<ѹ\J $T13X:%R$eolP%QFxzf[dI&D $ 4oɂ}Q gQ8i:N8Ddg(GO]AAdr@Y :.RS? "˪*En1ק*I0?CrRZw&bfV?(NSs]#PkgQSS`FF:8Otd |L*)) '&+6PPby`N =ELd!R؞m&hY P;*ٕ$)` `truNX)%+KXhb޳I{w7uQоfOnKz=*!Gh5l◔HF0)墴*O r<1 ;>|5u" wvjA유źX%vkꒌTsT"%A7Kϗ!mo/ݿ"[SQyTs_'-??<1l|֐ٞ|IBLy`f=Ho-/EɷT CLM"ARq;0#Za6k d %t$ &kZUZ77SoUxPA=6$yۭ"6*^tv6Z-&#5og!YPht%nJJA;o,^Dt5!Y[dqݣ7zcȂFhґUB4dTSvp& j;zaH%w/*O d #xZ Zs^Z(Զv=J;LDo7> !؀㤖`mR,Rs^` QblX0L`XYk m!rso^@҇c)Z}i2O]|Tu>{pLUew>֓r8ن<)#W3-^?[FOݪ_.Q-3wONv A4^@Gn-NA}L'GZRş'1% XQ=Ξ/ 왒o tKXG _jqz[ m\6Rhbeen޹_N0vr厨ڎK`.<ėǷ?A8eWG̩j2ޒKݧ!۱[4w>ƀ B1U5ʽxWO⃙~BPС ;+.Cq߉{-ʼn5Sɜ :gE_h(3A 0?_oKheAFdAY-km]&UkvY}:IS$\Ny}LIm {W+ U|؆q|wS~جCmb;?Tf ^-,Kzχ }ɼ2(g\_}«3`HgB*ccJ^͖5nٷk25 .pq69r`NN{363Ma9a<?7l?.l-\&!1{JP?F*S=%me4޳nO_ՔTS6ff-8H~vT(qQ:$9ˀo96dkZ9@H5o[C=Z5K)%8H%`ڣI6R)1l[7Z\\b;rpt`@9$'Hͯԝ9Ԗ5NSS y<zv`1>x:k2ngrjv|϶Nu8u[/?TYDKgK 6*Yl*l*z6/5Y\^h<.ώKAP>`;T+wlAiFCmgLçm$D궭# @yS|D\[ǻKB8ZK"ݪ9\[$۳MQ;Z;x8hywQYjpVv=.k\EP +iSI*S;O|]Q" );РwC[Qh)ᓑ{P(J:D^gnmYgFU}ֶۻӄ%ɢٻq%WyYb``epf3OIuLr濟v˶ԦDR{<HůŪb]$%#Ȕђ.sܖ b?b?#I+?z *oC-|uo.%s,EA W0b%7FgR LuI Wp(xo'&iT1#C&Fj*xIlX #p!_ "ՆqZ \ ^L% 'FګC:6-2ߣ`\`;u L-M9aVT+Lb*aIS,s G04r !HNJp͐;)DZfNzG!\. iQJ}'DqHr"^ YITJf-,(* AJqiiMpH˜QH ֕*}|apP㜁kKU`I mGpp33I܁:8kwDB'/|]透_(=՟f}(t^E3ud u}k&P8|l~zpf6G$(z1HCFu9gA~aw⏂\&j@GakU+_}f#rgW{j;7ѷ#WqH L3{wXN^ zx7aXl°™ 5 s~zL)' `|{61^q<(sig{=~tpÄ]KxqFv8 ݫկ߿9= `b55roRf; :Y.srEq#YT OS)\Jcq0l3)`Su3Yҫ߯D;wjb)&L;le)r-ګ2ܿyz2 BJM)IFRjMf撅`̀ZO}QolZלvdz_hqG^֖] ʳ!;CϏeQGKQ1z![r^*bk _0`TeP\ajAWk#in@4S-A~TkV@!dU}e*4BH[x%> ,&jXۧZչB jc\UF,r5cU5 ̼ ^A_}*_X38n]P :~Woͣ?]krZ Oo{XNr<x񲓕|[I0_neう[)a-=kQrD$?UU'j7[,!.TݢRH\D{˔F|=חW!n񥏹! a.`7PEj}[|QݾṵrE~Yuz>AVqXy1runX| [1^1x 3]=I oog"#U]r.#m_γwSvaе`iΦObg8{'>68I,1fa̛ytctְ" &pVpd|J}LrbHG68%u S]2q̿~9D~ma-5UHBT7E燫C{λ=,|3};-^o>Zyн|{_dlk1fA6{MQcKk3ͺAy'uM]Aװ%d I'+2^*7fRBofv}HL{ yA-a ҄Pm`31cK,tt7F BhLA M!PTgh~&~l9.HB$j\,ܞGCϜiO[ %r:z0 c eEjdG):G3,EHA״l(ƌ+]#l-JaFo/. rUWK1!zuQup?MP]cpq<.1A.td3r|w4ʚo>33semS^K+';E5te[UyÝ[ diwO fv:`XU(%chr}cՑz La2%S-̧S#YM8Ժ͉"R(l{˺1Jn.]* >j] :dJkMB=˨,E } X4f{XCDKi`'7lRk3}ݏV\%`ׂ4"O삥Ͽ9F9kVնӨ66 ˢ6AOѫ:H$\XZT5CVWr4*N"|;DΈ-Ljn11}25ߥѷ/%|xEђfV#^s_z*cuB>Wx_V]/^o7Wm˜d/D)Y|TUd[L"d&qKo BL4/&I83(%"ǒ-oLϧx~ :KObѡK"9F'0R^8Óo;zyvL l95+(%'^}Q4g,2{ gx7s8l\b=9rں/G.űBDyCCQ>\|o y_md_/gpϦ/a`?'k# Bk(W\~?f:֩ k 9qVߎnBg3KDfQj/l+Z2jJ~a g1')I- kc %N䧷w|bD갑ښDYrCkJИ3?cdQ&r -Ӝzl5xq"ōHG籍j#fCsδ3_vQX 3   OǏx+Zqv|C/>cxXNyaypf{| [K$ΙTQw4E"K)}\ݼT&%YRe\v. S9)IuzA nFǖG(&>_QTX|]9?jW-xSJ vQGaѓȼh/g26~?/-m~YD\ T#QDsI4 y$r62C'0h f"+[o[8ǐN8Y QzP;v4YXpc`FH t}w'd֛F(Dl.KQ%-yD}R-̼궻EL$jBuTbiDNB%,϶>!5Z|+2P-7@|]4fnC]4f  SHO5rb-FoO隑U?^|:'ږdgl]voo֞1Ma;z1>aSA~+m4o:;X'65HD9NJ4Dφ,-e9Uk9БH $N<>[9WbCD:[Fv dR%sqHM%օ)lE^޵5mdlvR/ÔI6SqI%[FwC"rӠ.KlsӧOW.{`k!ԯ\bĎa4C%LLb2GH"zyZ/Q0I^ M%U2aēxm={l3{k%V$w֖e 0~\qCqDIԍ74uX G,LX,K$(al;O6 UY, 3C<Ilx#5vȉyZrjpTC~E)XDҌic!` @$ȱ:߅ϊhsxz=_R-T%U]m\/pצ'1,X玷8I@DC iB'{K/ 6jW@i0T'0ݰ;/p\+; >3L:SGxJ3Y.` YX6bfkV/E*[;GjNB ^+ xj!*EcU'e;vtm 1@!\<^Nc- ؚpQ~8?!ü+!ilpoAVAI: a|c?΁`ZBtɹ8JBg/5D`V*Hbg'onȫVY[31>7  s@ك0 {3>?~z/ Axd_>?I\"9FWNjs4aW*s~xi2i!P?T9J [DY]eWkdSbE22B!;WxO9ϋ FQ6zba  >biaC &4{nx"`q c\= R=/&PTmUA<Q,V#|4b)kIq/F`dhP t Rdj81)ӊk0Mf\T3R"!̂Fkex7uRX?8{ fB0{٫^٢s,`NhMRoQ*)NpJ¼U!141"hzʢ]UrtU^EE(땘Ӣ?*rd<ۃUG߽^ĩ߆a1{?}8^񢺒Y BL ?Gr+c< e2C%YfS҉,͜ҞL`Qb+Շ%ji12'(@TqTj zE=61A*crS-/ci Lz ~Qk=sG+RL _5Ca;sZRl() ӧ$L0}JӧjN”m+X3S :3F!)=rdz PP i|UUB!~([… H *N{j0YF$ge֑l^8Yc2s3Y&FR$2'$ %Up&Ld³*h) 8xdo$tO^SE0oʧ0?D 57SM\͆ȁC(3gjR&2/@8'5KRB, 1mQJ~:kӻ#X E^F9[0q. Ac\ J7ҸrY <q0y8&&ory7qd=<$I8h!fN|dD%Y)TKasL79 !-BOrPQ30fpK;9[ @R )jKlVJ}Lн6SRy'9R8^j$˨u+ZI^l{\f"e8#`KfQ3 omYxX2 *35<Z!LZL%F5,ls]̰Jb*pT}ǂ M.5q\sS OTg!0`+ku#{h0nypFƟޙGãσJ?<{yoqF^؝4 Ð;0;7^+f231/Rivn[uf[ Q6f7xٺHY JJ"om̲RdkvvD+(u]4\\m[O J> % 'tk8ip^3r8o9sH3^8R7*.P`ub5BlZQjUE(m]Ͽu2׋QW= FwI/_-.m52 )FEnzPSz@P~hQڪ&^LQpj(TlgN$ f"߃~J"QQ!.&e 7$@ \ͱ>YCYjdz+pɫ4W9 gl吧 =ː~ʎNek0/edA+]Ìt&/8ULF#|:M,締ހy0 Rtj0M 5+W.èbh|?~q9/@os  BIo/n~^ Ȍ {7xyEz7+|E2͒t^VZ>ËQd^71}XZ|9|8AU UqfOpJi p36YBGrτ]}KVĜbt럭Z?|t'\@QRv9\ ^v6J0*//cj폴bRԵؘJ+KDFE߿/3N6U6L>/~+3?QNwW O 3- y1cp|3+As XWٖ#O`|>'z5@HɦMFj4GAfЦ $ξ񧜭M &H_T'lN0U1aL]I7׮=f2XCS.%+)M:(˨OnCga54GBGҾ{UԳh\?eL_ђQ2ݎ?GHmqg LلcL(DLs"űA"zX|\ V/oڴ!$G^-%0^g~H3έūK I9' kv+.-f1ۥuBn݇v9Kߟ1X/`ͦiLt,:Qicd,* _-Fwy% Jj s/6JY6)pkG]UR'ç &.:+1K! x ZBYQKCEvCҶ~GE=k<-pHkƩɚ'̉nQ?&w:{,}% X7^kfcy˷@k B?dذL{kMp$  i4i@ѶY|ލ1rZfCb )CL4V;t^Zx4jC2HD=;Kbd)qc#NH[U'4-Ë&zV{mY{Oewΐ;+xޅ{~D-_%oK=, g~a{fI'\gx~x.Ԗ6vRd Y >fZH\G}c^jt 4Oo))!BH y9.G0?H30Rc!fG1iss$%)k言].'ČSǺ֤QiW¢5=<&.8vR!sWGE EGPBcbH&!Z-<|񓀩+wFlxƱS-f5\} qpK&?$0q'oNhP喧5 kٝjS wd'n@ǸQe!. Fj/ S VRB]Y=]/V5h'r#5h'8RipyWCjMi#}t4ט 0:PHG#egqLI}+BvAqcrh$"gHbҵHw$gc NS?ޏ݂YVGC0|fٝ ?m8r.tnvzJp˞0 ~oeFTb:$5twfpE*]ko+?@>$nQ4\ZM`,wYeiΒк")'"gfgg2SHX[H (L0]<0dYįޞ<Ɂx!|X Jw)I?}Hlgjސ;7#Oh on| wfu_֬.^xz7߼zWӻgo^=/ :w{7^{v?p|s]ߞzCOz6z{+4{ϟ要PħL:wz?5ԓ ~~-^gfv/؛A\{_ƣ~iWg~lO]WO8+5ya4U!H~VSunf↟PS#zC}:Ol9W5SW!=]t 2E o4Ӵݝ7[/_l/ƏtʜxoOga$WA+ُ/ է>4 Pg; [g- Mu]4}}@_?~>~xZrԽ%͟@O4} H D ٿ2LA ϥ_ұ^oyVy 6S?7Bq'{["9b!KxOG׃QǯR__(4+PT^zP +s$4p"\g俿e$5BAg+]'岆L|,tݻ8DCS}9?E(eI$EQDs8Yfwib|̇5efN?LgG }s/M@UߎBsEڭYV681*&7>fJZE(uN54D>g;.]pj3A9'KH;҅4[`Su%L]$IP2$IgCvw5e-|lO뿪& ]E Qy;aY_|}GX69v}\*/WJB-_CjBmDZ6XKq]FZ/]3>OtMfvNb6 "J "@L}&'ǖF,Nlċ|cbɭVXC/S_[ÉqDj)n8A|tf)2\ڰ^!+6=2_X@B iTB7;0FKs*6i6ۄݓmhTn7(=4e bs$|ݚ2ahxLV?ݛM, 60vΕH7Ք§~)cؗOf4L~6HIdn |tʄ{!Z:q1gmm1܆cn]r+.DwNV3kaNc) ZQ^1aac GI(V\F>2ͩV2E.i1]Dp3nT^qXagHyGΚ1Ef\I*bFi[5;i-vs*v ۻXt{&iS2.'H"#m"Ĭbw fl&S_a u2#')(i?Š}b3^$rzl2[ V.ñⱕ,׉ qdb1$LZ0*beA{IR 8bˈ'O8sĔY,,BTiqb^-9e94-SHGqb-Fe2$,5p0((9$c'NTh`=.)5#F b]$ 72Ŗ:YBbq?%Ss3D#`PRfGbDk ӌܙMu01ӷ_.ú鞯0j^mJw `A /.o?E)<{|NΟO͟˞|M:,\#Db̙7gG0/i'GGٿBgٿ;=XeLPzuC@ 2 V;O't1ΏrG(.̡" Rr 5k*!fQZjHF2UZ rCVY~H%rkdjp$g˼d'Z")_VR!䪺M]QTs]:ץ` bupR8#qa0p0CO 1$YUࣔF2pA1C41Xf٤4XuO#A&1+xĆz1!m)fe^4 |a!"^Zpvke!:3W{Ӧf΢gi#=9L-IT ^o(x|%ЯD؅.W;D:UTޅ>1;Rls/T掚LZֳ}ݲӥF RТ q?H哥8ClemJ'qwO:I%Nn߷P\~η[d'<Ր;!DFL ̀x3&!8ZZ4'nF=XD\E==Xm6y-c9S9Δ)#qysqާaqދw7N~1BPp59Δ[!pbnKM@IS}5?ѴW[Ҽܲ@`ٵxvrEdz[uǻXK3xc🫗۵famڲm||t֚uVd7=]hzbnχrW?,g[YE@3§ל*ݹvgZ--Lmi{>>ܣ[ ja>W-`\Ss~FjyoHЕ =Baɋ~O(ⴕxrBWeqX˥]REIၮ'=49mɸsNށR L$Y t|lN)/"M)=`qM3t&> ! .JefΥo7_m.Nۑ N<;dͅ$C+ $NJDNK >&K`8^M!>E`dY/.Su[DHcͿċv 1C.|zͩ^ԒZ .}Cd])` W`eĤu@TY֯X kUa:tHwi^RTpnLFQGOH3U *Rg! VYy[e飽Uj b #B G~-VjuQj- 'w+WLRTD"!MR eL@Xi1s] aJ$l ;nmc>|Vàq!3 TUՉ I2cv7MlpHgRwtvN][QÜ& qܨ9\ֲ !@4k%вYйD`Ö!|Dƫ!D4(~>%D'+VNſ/h${σ KեHTE#!e4whe:1%_׳/z`EB ̎_b({ͮK} 7NycpXMfZ5RG{ trPVf]$0 [$LA&t:Rg<';w: v7lCb)\ 壯ȵ7\/}1 h5K`dfiqꬆl-O(@SgVR'3.+qv 62#bւB$G&D4r$ 2@ɁgPPԣ˳\cKAVr׵H`F֜v!H >%+&\[I9NK꠸K\P}T!w{aV9K9IaunCڞhK1,T6;kjEeyxhk/U ъhڋ FѴ6\ƏmnM pu4m0b(,tO~էR$9k/DD_cE9hAPZ |; '?vSk<  b @Np )y(|^\|lpxa?˥ًh8h{Im8L ʱ. NHVD%:i*4$HKHnJe%gǹ~N@KSo#r^[ ~ң3;L}_$gәyvBٴjW,Z8s7U7CWol8Kt 3r<&翷4#/^_YT0{ݼә$/KK0+yrxLOg0bv79N8'S_~eS (26,8Ge TET(Yx rvwn>}3dGUz-E>'cD,D@:.Uڶ+O,|6:]v% w+#Gq6V3t;=al4jU)HQ9+D! Hy T4I))G쨋%{6$"]ی(!}sHu&Uf(" b"M){%4 .ݼ8j%\Fٷ;gmĔ! #OIxF501g_ǩ;N3o3en2rWT Wz@̀G,Ae/x:ۛ1hS_ =G ? _һGNz4&6=!TA{IV:8 r"ps#83xwKr`TRc3zz7N󃏑hP 61 $k[m SH@ QXm_7-4=f š+ 2.`exMYuш=.+j fǹV:v7<.bBikGzes G=G{7zAAuTp$>RJ.|nf윓\;'(SE 9Y̟qA&2K%1P2U(0)* w&*G319UDMn{6mUh<[ #@Vh|JM;#)B2Sx8v 3/.ֳRZj%(SygHTASB BHY&׬hmEnR|>!BF`!5/S$5RyF (9PC҂f"M m|8}3 p Hj'ʍTI2@1̰pJpN1U Yf ?5&Dޣ ESVJ^Ket/On ="m 9,]eǛRp=vSߏQw1i'ujA(Zyo*n|6fh`*,myMh:^5^`omuzU>ާw5ʄJSClWrh6k[~a[wwytp;1 c{T9Y΍_֨ gI4 y;U^-dZ)V(S CdA(Bk ylBf1X 3͇ ӑ6èuOGg.f"Dĸ_ $ 9oɑ~tR3^)-DfBN7) ehM6:_F//9N@&'z&7LH$ I{&ǽpJszo:9hEu苰r ֔C$BO`CzzՑ֤UC1.2SֶHi.Y;5\ ;a7 Ȧxf^Ԗ{|'c^q` TJS^dyAR(oRy5)\ 壯ʵ7\/5x _GHE {(-!4X%#ڏ~HJ_7^4/oM'KXDUÝSiBƱ8sgOLϡ e2KrB02ndCbSXca % q /l-+)6n4?> ;~2DMĔƄs,urO!Ÿu94&og" ).r,=2* .߾w ,5Dž5DqӲJߧb)M4HUB ĝ @(%E'߁es4s3Fp9$h45憞, 6П9!P$6s,rH,Ghnl$$, `u.Or|0+-T| 8՘~м^Ȝrd Yϋ<$Ml=V4ι969i uQb8N?'Ɵ׿{ gǹ~ $1 j !БIQ$e.)U&jzVC> =&R9ǚ {I Q$ 8cYMfϜX(e Yn `HQ"RɰJ5⪻0LOMuтu]>\  d\'n:I+ {9ཽZa;1h][oH+_]}jyd3 &3/;Xgeّ$翟jIiID XfR5,16aUZD3V[^]LrTq>z78>)Aoeo?vr-08}3òR%o#i2(|H EEL;c9 w\dfo4gvk۶eyHոm[(\ j|Z5 m?cǨFu| iM݆ 9$S~{Ʋ{>˄F#s(qlԹV=IJFJ֚흴<^Zb&ZNϓ7&,Үy;;kmZ` x{e8vZO``|((T{U'@k4+8ĺ>Q`7|}i#d}#^:XK7/#& boÄ #ME|G|CD|<]+noL@BANRP&uuYLf;1o'k 2mBq\R(0OcN16]L*lqhعZm7%-"=4~~7Ax79 | "Br@ &7bL:u#D;M kqpF@,>T#j>_nYx.Gv~7DP xӿͯb_8ZơwVbɅ),UFǙZvnP[RfQgGi",O_/o5"Jk:k-rH5 yA9ԀW9Yo~mˋ 9%hWN^NFs;zbQb7פw(l^vg& Ȫ`$ݠx|1Y-?M«AE۞LڭcG2aM1YZCh^xw?A&V̜탚"%?D~SSDsNV&~Rzreԙ0TfP"9>Bj4:ϭ60pXΨYp1aeqkhyj MP(|J#W'qDPp h9'R KNAa@;,MYp26QtRǧNI[ a\*Q3s+OqZP e!]F'^hb:> D`=~@ {)QAk[i`☠{1)4X-* '(#9*YpbD0*Gh'JgiTf0!,Co1cX,jZru Ȝr-O!ƸAp`KiN:,5bڮ NwM:{F 35M>5ClXt)/~$5O<{`n >JZRS0:&wݍ?NF-61tILòWB;oGEao7*!w ;@/\pЕ7?lp)Jx"cCy’O79D)--'wek!7z+D(񶲩MEknjYak"w 6KSk֚ŴΒ{K E ۸ Y|P2%6yN5Ҭ%u-[AZELW$ԤJzRBMWQ,J՛,Nәp+sҘ,$`̯DqJ g^:}Z+d3=x" gzVGYcBTEHucdEq4V|Gȁ֢c[9cdFYwn?mGL=]Wb4rA^Wucc =E4;B{S*1m~01As͆VSj/ܙSx] S, xU; I~thUѿ8Ѡ~wC1`w 1:zpAl(ѿ_!LOfM]ץ?57?9%@Suדi%>J9J`+~ %UB 5OgjȝڝamèQ /dxcAM,K%/nFuĶ-Su:4#>  PMp%,I͉8(Xq??%;isHaVl44VYrΚ FbEQIs& lܾ 53nu&*=%FCp\@P4b+aCS! ,uH RЪqֲ0u 4(n D HmR 6 šnlA lD"mx\vJsgqSd8%AY!*E95MZq;RNU +rʉ70"BNbV cQ)o[:v{4 BK p4h(I >(bg.q|*--| ĴDEp:'}15ZEiѪMT܉ϵV)E rˌUB^ pgUv2iL)4hpч ÀSh߸JӴ/L>.q,TвPNr1I"4М17D+t8S:Xbu@E˭D.[C(bBfob KiO@=A?]x90ɈI =Cm}P>b[j94ET1n"\-|6i^72GGEO)Dkp1F+ɠnB^ 4U{1)JU 0+7b)1G,{sJn9agsLqB" '4P1h8V4M ǒN5E8 eⰋ/Ҙ'O~zgܼd7He4w"|$֘dWTQrmuWɻ\9iMiJ$ަgJXXy)fNZK0kF8seL.;MKWZi4ƚ m`OBwބ}\!ʒ> hhe [Č` H "3&@ UFa532up,4B[jjIZNi}O34~:?k4!iOC=Bd2(%Āf9:N/:ΒДZd\XDKha;MRt9͹%jRw~ڛ&T+t>ďF6~hpAp2m0Ji5܈i#4-)K{o*A5Qi \Ysm16~zw96P% $菞kA-T+ݕs;;H\K/9x}JU/ lL]aw> R4) ~^Ap&;xVF6a:PoR9@Pfz2_+1\ȹtk>5/N{~yjup t?1KQ.xlQFˎ~bQbC{'\mt>-W*Ҹܴ$\4;%2yPbKiLZXyLGӛLjy籏 h!^115MCpDMp9QZ# Ǹkd1qi4'< OlN GB N<7YV|zC3󣔏 Bä{ӛx~/wOwkkn痓;}}T/7YύM?zDZR*%%pCfb][ 7w+(zwBֵ8:\39M #JuXr]bPF\Etv$f*ߤ@~u_wZlY}]PKǶ4d|7Oi7ƈڈTFyϓDWi&4U7dP*_LO]׌$2a̸71>-DBD&zX> ˲uTn/{}Vl||1f1. *RGPp>:j!0m:}#*W?_REȒRу۫?E8}3Hs)Tf2adG9A0&\e< jVլO:0%<њD1 w%B@ tP 5n5Um%r||S9SЌb,iݰ!Dip-u|4ZK3mKØ麈il[1tu)[}!T")Xo^^xݚ6bK3ԃu|:ͷ<ŗ/nϯFvqf}cƍ|?s{k}ggl\^\N|wW{EH+F9e Jy!R^udksK@[eYt Rsa:DF8^O:1bi)erZ]#G*u JW)F)|:}QAՋYx:jQF"ʡ(*Ϋ|(lJE4P >5,gcjV~*!=ɜbnZIgzIn_]y,Nf}[Jdّd%vKjIZٱժbd,\)XFFS- e+RJigRbl0e 3s2o,ӭ@NiQrD|Ƃę:ϸ'KFvE "eEN92It8ڨljGuu5{P'C[8~F ޱ{tm(83C!DNnO?+ȁl˸M2~E3Z|#"tlt&B-FmnMuD:9#vYCNug91\fZf E,,ߔvoeSV84mD+wLeWB7׮D{;UpǨwxWZ08҆9 }mkI]f%_b^luKtB5chͫpaJ$R-}0ץßǷme٥G)_.XY1Zyh?˦;IXoϐm'SZ`I<%Z}րjFI%}Uщn=Wmʹ ʡ?X%g|oU`籂OL>6.`/0Y>mﴇ|?:8|=\/ϔr߃h? |=G1Xȶ9*ˏ;r]Lz~fDPPuw:a5cpe*xab?2&1^psV_I#Ѕ3`ϔͶO*F#.Ә DL"lm.*'IN`$ JI4KnnNt96oh0`u=Al>`v[N }kȱL٬4[c+҄y4̼AzB1d3ev2"Kh:[wz_`c`Q# ԯg,*0Mw8*4 G-!}4nIca>kN&h܌j3zՖ}#3v!-yͲ،z=匳U(oT=9 )JA-/ -F^rr 8Ab1b\ :Wj#~J [4ȶ _JfU|#94˝^ˀ, +j"•'#t\8 XIgEuR`) /q˜O ̼ YS(,b1p{n:մp̨֗*/_[b7|UMtީgAKr()eMaK͈A LsJy ?K=̹ R CKT ̝yjt, `5 l\ױ_#!Z06 Yy`tIY'E>"**\skG ^ZYx5\@f4[,$]fɔds`IJOZF۞^ 1U_q"F+ %32~/GAd?\FnQk.߬I/bCBY.8֫p;w Q~"oܬtx=;(=+@sОCԦK :?oڏU]ޅPG"lm YLs.kP ^O ]MZ9?vV :$F,>M9+ǒ8QE exF`DF/`!8e75\!q 8|8:POͱ;!QGT]8o']_ϥ:=_+A[x AUXZ8$q_[$eG3B%2B#DRU|{*_p݊N2;GuhDiWvbWbG{EAO 7Y2Fj'1OO\0L>u02mAe6%lL)tJi:3 9*=ȯIV3LIZHt%bRQ0nr3w|:ݛ͇G3 rꋘXtb9˃HzyϹRFy2A>,XӜlֲy3#rʼnS6FFGHy|S8*Or}sjB؞[ x7p$63Gz-7 }Y[AF`3+o$ ePKU^RSrrJDŽ&{k,rL"m^M@#Pƞ\ZY`5~8f<NKBEVB L5-UHф ~?yٶ%@I'\וGwK~ Ӈ 0w0_f f,Q&s)7f7&zv ʌ3w?xrI>ܧ0\j0bENPspSx0JPH B`*He2[,Lۜ8šzW m =+ѬG~G εp eAXm\`Hb{W眠B"q5b~50O:,h_5dX^x|뗧!>9YX3 ,GOCpNgADdf/7xd:[rh4_8IwFf: Mcx>XB[4v<]&Hc.aW eB.MH!eUxҤO5xb1W1sc1W1sUk:VF:Ƃ7`=V Îq ;])p"yr ޺ժ[wW+%F;5i"4!،╧^)ݧ^ bioT$mrqn-6X0.X+[z ># !{baf.5te6af7ig5ҝu}~s~9z bS*ح`[փM;oZփMzx[=صw}`Y*̋Z/ EeOwΫw1iMsH ]Q;g`9 5Q*Ӛ+qN.ysV^i͝>`v:f!FI/_]x\p#Dm8VU_Bt$mǕZEחDZEEԗvc5GUZC;_uh p[,0D8'c^nv DJ|YL sDsFI%0a&3yΑ֘)8\dN47@Sc fωb!48'")V q,@sɘ᠌9:(iN,Y&9y. Gw&j5sVx!1t7y#s%5"G:p\,RB@V( GqA089),ymz>_9#'fp=1. `^rZ[i- z^v%-`{rcxPz4^ )D;BgP^bAHBt/u^(m_\F&Ϣ|=~ k/ל#Yd0C$;WZ0hvj4T(pr]8L[@Vx ir4XOasKQ[ ldx]PTkF BhQR#֤<3Fun|K/F;5 bã$e g_S1^|?Xq̱phLĤzBɋDYty)K& 8TF'[{F7ϵĒj~9ثhK^qY:"BbPY"<>&Gv4jc724΀Bt!F5 EfYBrsqM R-ʸ_UB1v^qBuyIv(Ofߝ^ڨl}%Q-1:؏u*$=\/G1֥g 'aL[NŪOy) W^iNK)s<$ףQ0$[iKQ߽ [շۆ9mg$;- *Y_޶ $a'?9A8I6_h[s?z/\fZ?&9iqlkDtiM]Ëyԣ_wmmrjJ_\凔:Td_Ra0=4%)o )jxPbXph42̽B}4KL̘ΞI|yzٳBBpM<$XXY3xtg>?̿4l=,^ څ%E(W+} %}N]6FpWDl1쥹k="%f=¢/:L8_qK{ZCL_G8'icP7)݃;}K{nQюk)̝X-(Mq>E=I*߀ nʶ*u\@oe; f莔T%_jT2zŖP$h PHرr s8ސ}n_ȗk΄03 ƾ t|}hv"ٝYeWkbec z}\ AAoTYz4n_k^`'/%2(ʷs3]v%azd}d\?~Mف oIݓs@jm&F{YX<V$f# :d,y7yҞ1ovm耗w#/sgn.,rIVait:O2|ir?d|잊TZ)dh8:=g4#e>`l[㐅 $Uۑwk tgg pfd(dN- v}[8hb%@FitjmM": 3PXā/fEuu:՜Ur#qIV(i1Դ)9-DJbڋQ1g[ DP-[Lm-1==fKɞ|֊ۿ}SWd @{I55cKa#ˀ9[a`+Aj~Gwv򛙳ofǥ3:a|f3d/{>cA̪0F9ZkJxj+z~̍9iߨ~Qm9Bz*X3O$!\Ddp[ڍZ[,!;F֝xJ!iMEO4TW.2e]|S,i1d!w[Ȼ{hV_(/ԏçޝ|pPMs|W7_f<*nCyncJs*wD5jf22,s9guM!ݚ+Q RGtɈ2WDAмprmKܑ -x=oG3VZ)F0 Tw$0lYЈ; tɼ2=pL&1̀%jn&|_f1vqk˧b +~Oox/q2q{wA՚ ;sHt㮗 *7 w,)(%*ҔFhw|%׺Ē1JJ%.u-%39g!JvJ(˂Bc ) + 4Y pw,FEZ" :OYxejr=Ā+!˵_ԄF 7=̏^Kɘh6-X6z\QA$wo^Fw)%u*{;d:mپ&SpR0NZ7E & Cj#`92[c'*6/2=^>65 1-c]t25}y]ڌy<_{"}\ Ќ)]#Xb4%ktWHRF^bz, )aCrAn٢nm =i%\uջbZWR@{5ZN_5ZOdQkr)C,G㬐#k4,D>pi[gms1h"zEEC,dcS= ;ՌsKkde)g 5ĄQ(5g+' ”(ʊ ޢ%Ӫ_'8܁9V5)Z@8ԔPq"H cDXZ " VQhUL|[StpdfpQ+ Q H(,!z{z UfP!~>P%/ OQcbZ2tFqR"O,"Pbl]Z"1 `T8qnF0n7墔HX) 8 n:hGN̹;G5h}1T DzYxFwK$F&jvjj=T;L $5v&m-;:ޠ LN!!\Dd[M-;nNM[lD0PW.12=#pzv}>~%8N`p.{Bҕ@Ņ2]YTdqZho޿EMV$S? %_r?8mfg3wBV:'`E QܨANF3f]d4bEѐcң,] QKUӯu^`Jv|{?HͤSH֛DNO,.y{u{Xp?~]1ٱӜ3)~Az37ã]x @\u 7|&zv ,|$w4DƟ1JsXG:R):YНi?~}"P x6 tr R`Đ%JHV`0t冼Jͣoz1La{0CDKEl3ͳZ`YSDܪ@"/YN[]:cU}'gy%a[0mu1z K}ba-XvyD:b-N+["AϸrֹpHiK/W,d/M%n-F=U㡣a2IYdծ沃I«]1az\_/\tΊ'/ to'5HWޟzⲻ['$j}Kmnt<@ t3T9:dHHr꓈NJ^C%'6vN5@uMRMtDJxRVAHFt᪰LSF )9dZ%Ҋ0%F6k_uE5݇wL4\c4i&ڝTEO4q\ y"#S{c11hƞ[DC[ y"%SmFy/lX BD'v&ڭãi-N;vn)$䕋hLFk^i1jRE5 wv& ÿoo:[n&2bhYT! 2`%Х=cl<ĵ|W.6!#89QA,8X#bS?PS]aoU-ҼݾMUQЭcТb9!38e 8F}}wK!vDb1>ΰ[Ojzy.؂PszV%~V,zVGjM:nguwg1S7-vghыvL|bAL;/n+#iuh܍v.%ڈ`ZKh=y S6o,C?y c$tSȤVb)edX×ܳftUzpXO;ג՗2S 3Q"0+4jِZlȱEACne,6B777R (R-(kmH9}y $/Y}tBQH.j!{8+9=35_UWWUWUS\$L4k@-#!sJ/)jDJg]CG=QZY&({X;11(B6.7HSdt{XK4sÂSהTwyYw 9x:M'S0 "Ѐ7Es0HŸoɹ~Kn?|"ub'/KaťW*R`'HF"2+ D%X!")M:h)XpZrkn[Fe֜HHGJGI&Z䜀 =^'#d |tg/Ft1#W\\38y[oAO?<#`,j?_~~{*ah12~X q~G(MSDyq4F%LrO30') +,I+ym2xMw#3x!~P =\^hś O$øO 0늉s%AXם+ޟlƑ9 K8SJ!zRhM;*) i,fTWڒy-E'Eʲ($<;8YQc o\1 D~ =NXj=]^0we\ɉ_97#Gf؍tb=V}۸49BS )GVZ܍E{~-vsī#UװWa lκR3@&B\ ~.\lNu)6*mG9JtRj(DG¬q`3f9e p˱rnP_,~<gAo$CřYVm6F+(sɱd@.ǒ5\dKqd(LVBsuBL~45 NK)sUl@VR` R!^2R0DžR .P7hsbݶJS8u& 0[<:HsN-WǕn]u&8"V2XʈhTJ,HD2[ˣ!"Na? [ \t} +8m=M'W0Q:7жJl])05W 3vjă0s1r趛9yE9WƼ06{Y0oiM:CV]Wu]Lj5¸ ՆQ"ygH¹:q(+;Cy,7a+D1X`}mktrS0&>96/ 3Kx h*271P9`~np #pTqSKzgӷlޭ2LcK vqrc vu Q",v+tjqIU6Mig6h~U 03 l&"ENFnzAC## a+v 0?.1E?ۻ/궫PM1'eLPĎɤY*͚6km0vWfC@i/ƀ Jax{ fJ0 ,!G6ްjCFj0H0Fc(X bV{,2ODGaJhdHu@֨bK._7Y97ONֹΒRBGf4ұ(ia\kRDmu4l:d9-aG QyI2 |A!ȷ!>YE;$ NK4 d2(bnH>RJ# 9{*|6K+Ql`+eF: F"6`a'>3::#؞LI|L]mahoh]4:EڇyAawa—h3׳_ⱳ^7};A8b{SLv}nvįV`qiPm!W`mPk%qp^ѡve?O'ԋjEy540r\M8` #!=9_kx20pW_E wO]~ rg0Q Z1~.۳AyO.̟<_q1;@b|Ќ'/4r[i:D&ڒ.Wn/,襄J?|5w!寮㆛c@ #j^$]VNqUfu{ [~B5tl2at<wE af PO˳{!mYu9Fgw7z&i'ӗ 6 \싦ϘS^"߅f<9O\)Chiyʧ\hg~2y\~=Ff1?/%7V;(ݪaRbu,Vppkԁ k"Zry|*pTE%[ %ы]".Mܒ9m\scJn,pSok]}z ^ y7RގGX\I9'= ä($Z| p2,MMNaźxs9:ǂ\h`kpVomCx |[ހvB?|T櫹v><ԌdEj8TߑkdTnن^ֻ~[y6@1ooUrC 'GƧZI\q)#+&q1NedZ{IBx)::Ÿgjjj0JZؑ5"ٖLC4`Nܽk)8Z3iܑtJPL_C@is|0qKLH?kG#`>cxcpCnHkh G#kU*biҦ&Ƴo|5Ei7Œ`)A°q!TD8uҦ`H.7!lWӥ({Zn8PHfWy(׎jŭ!"([#eA NIX3(fe[7kׄjV9,--Xbݖ5砖v#LFRCT9Q r:gHN`B [Db:jɓ}>J=:]^yRӑ[ɼq- A PRCHOG!cC^+}D9$#l`ƣTU; c`pE|͠c4QcԊ X#Pyϟ6l@@)6uMZ,vY/k gS RQAGab;PDeQx)ϑX{, &ƺUITs+ewPEU샢Xŀ5k:fX,Z5㦛JÕ&@G SpEH{ISz̫yq<{"iݷ`R%̸i87uw7˳7_Û}xY7g5xy@Qy-|<)* a~N.t<,) %sRBn:HSUo>{w3rIkuN+Ǎ-'H >> ? X8pGxZVvH3tiU{Hge2E U"*kpU ^1#q{0Ҭb{;Mη 4Y o8݆]|4;/7VQ ;IB>'e7 XkFБzPM7\"{u kGSX]N_]õjuluR %*UC`Vk5>|>@D $smVzbH(1J.P2j'|m[ m6NC\ZfR-B&$BP Td,nu^ Gk>,8Q"m?jLZEPZ-,a? hVs8 CYOK2 a}0C).0(kmF/]<|? ̇ ],;\_ |xt'[-O2~dEI-nbnU,VPPqt5AxdVE du@($HM<P'BKVܻ 6/C3YAa?Kxr b_-{ %v(%~z|oK^]'d@ύ~"HztD:ō)D˷i7Owc5&+',;Qyx{kڍYuq=xcQi말2%7w^|V2@MX!mVxUL 6f\' 5;594xCQSz2A ,DW^( -7Xa G )Kh-SX)wz`a\#R:_s\6 Ͻ$mN`+IIm[%BgAnݻn)v2@t_FZ?456d+@n]@UjKh (7j?MWS߼JyrGi젤Mvۙa2Btߌ }]~Tݨi2*"Z]5cUyL )}a)aᇇ{F=+XϿcDekXq4#QP<^Pz#Fɒ~"pU$ϞTp`Àr*p%5{. 1AaM´E!;·#"7Toy!B ,vw ".V/:yMx$<:?7ʻ{ݏfEioh͞`l]%50q(,o~xOF M[HyÁz u&/;yFheli⾄sEY3XZ c/cI1Xa 3f%5MyskldžzI^n O$`LiĎcY4 ? /:b¥OI7aLd,G'xc\%6Ŏ1l*g3TSPT wLW3'#c~W,⫁;Wk0@=W;@MOceTg;:iZBo̶2<8b\Ul &Ǒ:@|5 ݟ.Wcpn𫥝8Ʊ B%"Cn@іR2 %ý˘s'Dbv:48*(sU3>4/.;< UBNX3L'UB)Ǘb:yJLlPhоǰn˶( g3CnL,S:C,lW]yWݨ z \|<+ ))xސ>מ^:Y,{/! 9z-Gz"sr#PLO~8C';~/?tA8wDx1DFDF =PPL:3 o?4)З]-RYE ЮB\_!W-=ѪU+=>Fb0Z(ПD[FDAćjbFJ }z`;_LXZfԯ)Nkv=;:S+>@2pt H&M'O_d儞)%n DqtvQ2SGJ9i͸1 QqE B DXA4TF"C:R%g4;"!!ƹtiv ib[e F@@c tz0Z^OO'uƒ வ. @@;$w7Oxz=͚⍂)c,.]!;Ezv.6&x{O{ #ez >m$(ӆ86cBPLq$2wЋÔWڇqP=cKݓr``e#$F;80SRxx^=6SP@D5!P!D*A8?wXK򦉥&X.ГXBьͲ.].}jfˮQ] l`e ^n!g5Ǫ2*@$uNȈXB[x"`q('D-9HeHU)=Q;h5& $[!NVw4tw S~*$$IVZɼLU[nCUKy$Ř^0fò%^״t37ӗ3/61?2[B4$vV1q% ֖y,+̙k' $c cءiX iz2,5v}wg0zOwDjS diz1q dכRo~yjc{2Սq:7ZXM1vQi2Oj^}{OHyEz18$Xho2 =_'GQѾ3@]f9P6kt U|eAI@p4bAjZ[e@Pj"5œ*':$L="`YPKa}kێɠnorj37^ٟĮSpgk]b!_.Xw r4,ggy >f)k˺d[yf[uv<^s QZTgͩf̃ҌlK9Ղ%j}G5RQgqyQmV5 ~e $#<=Q̞W?f+g׳Kt~O_ft4c,( /73\~d<c{ơ!wMG[}UY0r{N^~eRJ1?2{!= Q םG\G$!\Dsdʏ4oO1[.);FvF޴[~2EvkCB^|ڍEϨr1H1h3>A-(Q!!\Dsd*n2TzO [.);Fv3M_&j6$䕋hL! .SM\hZ] Ë^q?[] v_+/6pywp~B{+H詢q$cD-5l 4Db#G w 'Nbu8B@i. Dd0%ַѷ!uߊt;Z旇&YA0ٻFcWPꮾ998O'XUR•cO ICqHΰgh%_UWWue_?ocBI=~K4mh/*LtqoBRRlrzZYp4Bl/fuEa ٣~fg;wP v^>~OCڪgmǥyZpZ\zjIpQ%"K(ŏ\ĩ b$x"uI$e%Gl-FhI -+R@ A PftE.x+ar:E, ER, [hh+i@1OZ5Wy9z5UhO^w B .aÃx@ykPGf>ql"XZPE;Д^Debr- 8% 3 :jԁPсc &Д &)M& jf !fjKR߿gʢ'EϮiڷWO_祛zX}W*^ 84Ow|UX|;ܺK '(kD^_|ߨ+BARd_?5!  S4Ӫ[queq'a}\$u{X=X hԲ>ťFiTrg\VTExjz\*y)({_{';t>z_BQio܆j9le"sT]YXg%uQ]w|o!]y$ov{L!>6W =J-[]))]my°m[_,^>I%/U%TZ+POԆjP|)2tAB:4AܡmO7D9aC 1e]j'< l6Dš>Hh5 .!Z6N>V&wN^?99û:'7uƇ['7%00nNn[O- \Cv^4>@ІdЋɌzݶa`\Oط_Qiymrk8z÷/β1i9v9ĄO^oISA dA@OAEEKRN[`q"G>4jѷPz*I GS0B|j;!T쌤\P v\/lC5l:Wfu^z"oOHG.xc6$ێ )v.w4-xFqP ^r4`LESDtvY`'4=t oN773'a׃'K^P]RM 7c"ӣTY4Z_q6/p/qbPomŝ-xZEukњېw.92U>qǺ'nʝ9FvS\8>6n$䝋hLI1c-P|B햋Av;QF/ kj6$䝋hL 2ӏmVn{B햋Av;T޴[~@Mֆs͒)}v!\20|<h\ <ZߏϷ<0I)&zF߯I)\n%/Ohc%p'p'~B+Aq~~7~''>ώF?] Qy~~~':C?Ah*F?aZ] dF?͕ $c,'89C?A>a\ E -$QZ~~z7!HZ&2Y,kq ~芦7qRj|f!S?esV|Ѧiǃ+ +f5~]6Q>\|gTKGݓΥI I8U]pNeMh]ts*v{0cqWB/MP mJUH2\X3a1̂֞f쒜zZm)*tàvwȍjO/ YV(RmB%5e(M_L]|,'hɸo\(4YSwy]y,Wɡ3Ҍ?\ 8K',)3 qI{aMāϓjb XP.:uFzBqxpCDf+h5\%QU'o?i%OF 9XĵQѾ  /g/Y c M|JFԴ%uiGv6Juhڰ~x0] 5 Tt߻m]+҈n rRBsX]|D@UAuܿ&ZF=[Z†ځsܯ 3Ԗttsl]rGwdA9-Nt&pE.PTPf%L8~x=.^оz v~]Oj↼VPj'/ gtjŮT$>b2bfXk=ŷEÇ_ dfj}t>G`%\Xd72$z@O, ) Y1'd.6 kJ+rE΀h[?8x}cZq˧ ]PO81e3L9, _v0#Vz7I`8:ӎ3/DHG,wUd[kRO*yK ]A)=AMFCZj2cBӼ RV}tNzS+J"1+CmKqK@P$EBHh.uZh h-TT·Z-}|'{€PU07kUv{)P/JmD&dЄofi  #wF*ff}D}:̝dn0qW#(> n= X >PrkyyjЎKJFM>9)D!ḗ& ֡Rk[(iUr9O}`$gPq^d52(`f$Qex-e-h@xk(K](M*LFϝOyl+_?)yG8PU*}6 Suj齃VΠ={&3ߞlzV c rCͮj!}=4DLPa4#)aj̧Ogn4 !4,;}G<&u1`X9C[;ׯXwh]_]C\:q020n>NF-GQOd/lX` DeOC/6lp5V+*"@BwCNn-͓Sm1kA˷<&5G_c=l<٢7}S'SHmq|Bc( ]غ7D3&B*qvD/P5BFƸea/4GiK@;-q0Ь3w#]ʍ 9_$MڶMhI޵q$З!7J{v#Nv?-~ڌ%JGQ&iia #LĿןJ$I*M>|!cƳ> Y+-^RW!ۺ"kRRղS C4?6bZ9c~s.3 "paԀH8vңCtz]Do͝;{rL+עz\s_ t[؇?b; Ċ!+P=n)XcCOn,y~=٦шige'c ;'H{F.wM@0!$ӡ)Ugh'R53j Zyh0i~6g39Y8ʹb'@UzŰ4:2^zH t4G3tGs'fiYI;m/'.peDwlo>[YD4 <)53nҙ~]EA'^N`&(igp' x޳YSJukgڎwryUmp .(Pΰڬ}VR0z0ug?3Z-nCJ_,6zכzxഠZUGJSZsg4⭡IҜ6 cԆϰ-yCB"Rm ܏4T!?8uʯ-nhg41yUmrcY;ٍ'2{Jj6j-1JO>L?./GÂpnNO_\2W_?s2C>e|-7w//"&G!g/?\=- 7۾C.Rj*ӷG7/ÌK Lѵuiy[tykAS%??L&4ma UC|\JI<?e?+@BV| 'Ȱl7fb-e _ stJR NKv 4yKD؅+ "O*kh*5ڸJ66O=bJ1%,|Ds"2J/M0G6<8psv tz;Pj{#X\igo'/;Gah|B-o5Œ˓WD*P:a1*S4u7 Y4J'ͯIt~ҼBzm0Q'p#XD`35΂0lR"ReA= f\)rs YzTs"EjX0#EE5+k#,J iދjѹ~il/o&;^e4XW@}$^~]1^Y"<&$2,HbD $rAo쑞_k䉲muLd[}d>8O%L Q+)&4k\&:1eVK P #{%B T4D )D.W&ȓ.6Ƶx>Lb)_`\`ۼз/B2;P`Z6n*LWBgrB(B Z& SQb!GCrNNjLuza9baU(a%D-w0;vhA@hWeq;J GZ-@8 0O" w2V+dS][gzrK!KRIEȩfHgkcQvJcri [7PYLD20 G(`b``S7ۓ5 vmi56' -,j|B/ٚ:;XU2뙬"P7K [2oxa:gdM@S@tİEadAnU<塚L):+Dx+b(EJygOi#m^a+ C>c1_N#^;V"@Y&[ڻ Tͽ*;;+McV\";Y&Pb9,C\O𯱖u1J8Vk#ϱ*;ȎеdyQ3'}~fLQԎ^)oEP0Y >$S?F{- iķ$;N( -%%j%%>Xtz} P ʕ+9u(]sL5࢘C#ܒ()I'`RL&b ɾ6bҩ#f`QQK 6qkpQގ`#Lg _bLW%24c6ϲ3(L'~f3p 78SyV~ Mo?iE4JV]GŠԾ^VZcwpU+c8inog:g^7+y}"kcc,6܇%/JzGZ 7HPlyBÛ_I:89^l 4f, ,I[mD+Dߛߦh*нK]C\0wh †e.tVDl.}MO]('Z=[)h'X#`E1[t:9`m~ek  ?٭CZ^3ɡKAd2 CWځ9(3Jo_ypHSn;3QJ!ϊH HԒ>,6l'2`W!o=XrpN`]u ;OKյ.958 2NQ.u)t\[|,ֈEYh4fau,ɹi<~Sx-nE̱xv%|&"V۳)Bv+ų+T49_nwH ;ٕLNjd[$?Iv{5D֭/]@NR Bi$-ױt20h= iG=9x D !hag% ST"qNϵDܚ';(]gPOu٥rޛ e5,y= GkDLDF F ;6qvW]C8%4-x 3v1)zR0kxIJ"`&ѬžY_^4uZM>|%w r :)ːiT]0%$JZ%8ZxCp(׸*clU Y:V1 pPYxk~V̯'sZXxaژy,{(k!:CbǓ+2&۞>`[wmcmzf6~/z;, wi-m5pi^,̫ͫ[YSSl՚Twod!뭐hFvr BG{;S&f;G"F[ײ.-$:@"'m\ylYhRJCpaw0_vkLb>Cǻ١CI\- 1ןFa6,nE0mc xAC,$SD;t(tvA*6ӵft2wz˒Nn }<vT)pZ˥M5=hi,ucʤq$7J 1U]+x `RYWۇ `ݍݿl:5191l KW6SNt q ^2`& RY ݕ xXϟhlx;?̠ 0#Xaƪi:;GvU`Wn٭hrROdG@vDunTnƞNh j]5M԰&1җ0tH&+x%V4"pvs:-"*. ;f.3|KcT@wlaYjpɏhT2LS;ZuDuvla=jR-3;N+H{#-Y1r_}*1gs;ip"G8V’Yh-yVrp8hL]O<Qױ.'J<]x% )Usџ )isհhfE^Gh1FZƏYS1xbEg(L{oS ;`Rl.)d ;_dh_fg̾}-Wa' am z{ti(pbzoinp=  ƓyPߓjh?/~XuAaBov 5{x( ygs!e|5a:,>?FD0\"*=\v!= Q VŠ5#!?TDv&G5\-A2*FSQGQBfL'O4V7!!?&ɔ$`e".V Ipu!r-"#F=F<#l80]ATU㻟.9 cbƘ*eO)kHsvօ;$]tE.#*i{U'@s Qn kUwu\ׄ{;z Zs5ˇN~P\Q+=wG[<9}9 ^UYSLtc%50PY #S/pU0RCS:/"AF7S$`ʰ܅ U2l / M\1l\9)CJ7W/zb$K=hypg#vP~0)VL./ӫѰZ**f/&/ O>KK,|\NK!}RPdu3iiŰr3 q΃ ~b Isޑ.QOow_jGI{g(rI9Di۾esewT҂*N={v>f~㛕Ek@.wB]rYmYñ*d4U2ak %fp<8RJkb ^Pr_X ӆ]=3Xx!0 Fu9hú"fVOq_f+m̍K]NbO 7P\2ˑ9j5`^3 y48( 2]+&6OcL $0leve ђ^-"1,Ҫ | _\a>~>, Of%Y/*!ކo;WͪBk?KPI^_8+zS~g,Ob3Yx /RZ+TLK(|79(D1O8o&&B- Ybo7.'Y# è;5,WҹL::2#,+~N-f]wj2r2h]!ekݻWݓoyfhB6} G3[d3#&ٜ.P Jj_іoVM|*oշV"abR+|sz_>5`?r`]f]5qLBM|T ͏ AIh滕עYR 4zi; VYW@,Œ5ܬ~N,0Yَw }cd>+zéGg;BY9yLVU>֤:Z9(l 'Ph8r?}^q$⽨.ֲzR6cګ%{Κ?)6^d̂,XǨ<=y %[.s޳TVNZ+C>L}CDNz2hKT$T*\3:) $o-^%{H} H/%D=5 {3FXW;kjQu#b.S|8eܗ#"Vʸʶz5vXĹAw^Rӌ#Rvjqzn=Nn[W[ ''XuL@dvJ_DS4evnRGgv.h*31VQ<}Xsݫo HWO<-v-4tB;qMZx'J NǤV: KcfOBk#\%Wa[=ϹŸXj% c1%-!aمH^^}+\¾A23$\F OO 7; $?/܄pn @'Q޻"pC,ϵA9l^k5sd5WL|>AoПoIygo׋HSy?oN?LC+f_;/yM~Yx@ 4\&"kP2P!TD-Eܩjj+4oAC>4_f@朜"*<*9&\z) ;{DDx!ńJ PUj\\p2j0H%2BiLXdLբԽ8ETv٫%mdN3{0pG-j9(s4E Ri&pE:LCG~A5 $7]JGVCB~BȊGaNS.}|v=6Ja DWvȘo+f1_O(?bin.U!QY XO8֠r]kL(>~ch$}֯&\6]f!V5~E&0AR GЄ1< T w@Y (ĠBso9Fg('8dD;؜幱{ n`FeO!YRϘv1y%DCQ"x HCF+KX!Ԟ9ı*i !*8-|f;Z^uB-+\5Sh_z]Q.?E5 EuohA4$_*Ɖ`]7GTGG9RS1;}O#Gs$77LL\ƒ1DNԇlӂ9b u-0w(l` 2@T*TYn~N0sƄbiPbR),UM`7iZ"/HBbߖV}X _)A\pѧߓ?%PrTG Rw!Uoo/~8͗0 Ƿ۾ ЄBӷw?a&ĥB`,P'CXχ&H\ &K}a0iÈӍ/`L_,|ݭjM칪P h&LRuJ+ q CõdpQ-#ø8 xHR?n,Zhiښ!M5OMUJ ʘUk+aD \CmԆDP.ujC]b*9TF u1R 7d q7c*N6 A"P+x ]Yo#ɑ+^Qȫ/3}X6<^hU#Ψn}#K")e% B%UQ_DƑK鉔=7nɉPzpUdTw;'\P{̽q"#FiKz[ǠLKm& l?S VM򬅶1K˧/LpO ;PgtcJ칇Ϋnl~?巽}O4ŕ?XrdLOeuR{xrs ;QNyd'hZ4aRw1tO碡?DlobºY]+j%ZD3FZ  +9$`5NK.Ju($|NN!3vOp}׾^,8Z[=RsY|UoZ4M2 sS[޳M(Q 4vx!qhƱrs_޲_"SE%CW#Z+$RHJfRZqt*[0BvQ[evcOs/Hq#Ux&| A}'B9#f.aLADZ@7$ʖo j%L,y( <4P^ ŊzwgV5ʅ̙?bɒI)R*zEAjfbCH]U-~Iϗ.)^Uu}\/_~Pm(EΨ>Ћmd-3;yZw#. CЀ&7ldhuX5Ȓ ?#Fh&[~#8=e.pbuy4VĢ8nFIrp{}␊?O 'F;9ir&z޹;iߚ>i-VOZ yTm,|<'*0 Je9#d3 }?.;6[h~7 bL;_{%cȏY"?6%a1V': mE%J%Iz}%DK*?هz#>Juѻُ2$gu-\n$X9[]M5_]͟HGa8!AXػXDVR$VL0% 䢻ݾ^|"k 0aTeԞ 4NVP- ]4L)\{yPauT3}qZc(O+DP=Jh,l`3MһǙ\悞^Z+&ݓLw=]F5V<#$<%SCTFc~bhkVLeF]~p[d֨iN8#mm[صksOP3ŽqBpZ! 0#"j'iӢʓ{u+4u_6Jzڙs.KE?+MAiBoݷQo{IVBX^mRI>twḾ߂NnBՂ\%zrj9j_vl̢Z#A&T*z[WPeE4 Ws$KuIc3kT~XvAf ] ^2\ 3K/zV%SWKӬgl3ɚk1G P@/S) )K HGAW*&&jjn\d$QQNCHY`T?`ʦ:9\#"n]PcNQ|y`FxBiV ~I[zü݊7gChm CjQ?[$W$egYՋ?/i҅" nbY\嗳ޑkY.WRh&FRNZ*>^r GMś]kT:^T'JIx{|)Xߨ~Qy-<!>,;fᶠ?ALCmRpi.?}%S+R{^r*T1U\ӚPǖI;lxL%7|[R齃/F^xcB\ȤG̫1;F!\Ġ#', 3!e#Y,^B&dJ&_i S+~cI2dHXLJ=8ȹYISܞsCU V,KeK'fli7,icT`C-SBpHo%fxr^%eށD W6^GRXh22M-v(]fVHkYj$ >EIƀ#PZ֦|I{QcJDea\rO8(QXs1Fh'յp`D' j? EJMsHHIzBOEkJ3ܪ#S^tRj X+xXp$A4"V8U*yE1jђob7?d|GW᷏A֞.0$r^ݻg5nk<~_/_?ka~On3qU1Q5:٬/;NHzFB1n̥[4//?s}nm.,-}zA<I@ڳٿ*\8n3zOQ3)*뤞;A!_/SKkُi?s|G(0ydj+R;nD+/VL6A򎨿YfAs"VЧ=uv剄+ ffgHeH VW,[5YXy.z"wǪ$-33r95y2좪HF(xFn2vCPs+_JA@ցc5 1IP 剓ƺ.F[p3p^dS߷x@Aݐ}k*]`&s!2:0KpSFYaS\|uLY4c ? ` (3kriS]ZxXbI6F_aE-l-=,qEI)6BRʐXh_z -J =SL.L HzS='LeG5*;@&ְ|&374Hm[V\hRZǪ&TE$Bq.%XWNV |dA)ڠ\:Q%=7*gRtn'{/Z0S%*x;v(NJ ,9O.>AGc *A*'b z襖%0bmCj>mhX7ytjhPh@ u!:7Iy$EZ2+.,wD'^-T-mPRMIFѾxfGfΩHڇ.Sh %#kKEgM9F\jVM Lm+|-(g^SMٺF"[< աVAdU/J/9F+.*K]X.pdv@ T/E$5\ IoȏHUҵ\G-ӓG)Բ&o%!=}#dAq9X ct|ϮWUVSϺux|}l'N'%Ik-8kZ은"b0*N`-Xm9@*K]9g"EE.V Ѣm[Iռ֡uU[q6<%pf6i"iloHIQRxDͯ}("`gG~vŌ3 58GpfXv~x8' B4TiW.\]4C.[q´@80ύlENh-$\E`JOvTI!P*tPSD0Q*oTov>wu[ƞN8R߾6m1ijB0U+@$v*GGS!ǪrU0695t<9ŠASYSP i68H@W1eIzm<[5(HS%b)&I) QĽYPD*)#,E)2x! گI8 5_H"ZKܪ1)ᰴPDpjc'[UJ p[/lb*K3!972 RWX lLciiFΪc[aۂ[9FH󄹮m;b&XIQW5hLbL˛ŧyU@TXe1{o UuxÍfN#c:y?{@bonOfOQp9CY+RFX|P7~ptCpt-]38ҘC{U7OA5W 0D5g"% i^Ȳ Am|=N!4`d, [FzAe0ႷwapvF N{Cvc`'rz_ne핟jÍhb_uyYNmge%&8XIg2abR%G&z%Ur/SUıs܎E z:1/0!VO1Hx͔%V$&!<1ф:fLڿx[_wm`|ݿ;TKHXQ;ť XsbL,2mon5TOGc?܏mĄ9. -*]oqcr"Q8=]a}X@eӨ !hC隭S.1U++XGBz. *uAX:D$wLI2TrC.2ə5&73Є!eT  ^KglV*y.3f(3Ɂ!b`vLD2tnj!hMʫUͩb֩yjєdժ@#FUiuja7mT +ZU)„bmU{8(FS-AʵtU4 Ź-vKKr7kbGƛs5!!P]Sv\b) *^~ah'L#DwNOu՝Y s)˂2LeKy:ݴ2܂zZC>9>hE&ߓgIߞ2!:">ښ,1,3 M>gIaisAj /er*$cNy*1F"$:.%l!5)g4NaDCT0FJ"FZ hoYIQD^@Xj|ĥcĴLvHՇC{&+&nM;ޫ1?Od^1˦H[X!~mOp{Ի;!tyIȐ6qљk0~0[,Ba}Ư'?_$%,h5y}8UW[ipan'擹{Hw~z5CrX 1SA-?k9O~d~ѵ{#z~c7lW*=O^\A7{ąIL$oRvy_M@%׻kL>fd0~<ҕ3nPp}@F5T*}U7&l 1J^ӝ wݩF1ZiR|B!<7wOj=|b f{,~5|;QGt~p};=S'hܺhh^ec`瘷[)߱_ 5`ܤfG dlxT LϖMQF(W#ް]}p9bL9B}5m}M1"䱺:ou(}րi1C``2[ߛ 884^xhJK~馸]oWţ+  Bҝ-YDa9 U<3UՙwVYuiGk4SWǃ3k3$#3xϳx &n7жMi*`Z6Л/Ljx j2Fi+1[A9̢\U.gs7knjs^8/EA\w3ڜs4Bcj߱lq݋ؖ/!wPz;ƈ4fl$u,a1SUg{bW7ֲ=\W~Ϋ޾qF  ̭,RJEG-+H|p*08˽@I7gym~Д^2Ӵ{}}:zsE!^O _ ZuO04m+=#ev8xz`3dviP]m62fv Dz3evqJ'!J=<9<מ04GR-䄯 pw_/)WdjDgt@% ͼЈT;+U&U^fFi 90HAv4ٕ}{ꮗM f=W_KFJ EoԅRq9A3 */)M, 3@OjN` +gD{+sc*řff#tA)\jZa0 eA͌,?5Vr0f^ 0YQvR[= 8>VדyvFbżT\S&;mzv~2hre- {CDPEF'W8"z\,<))~H-)^h(;IND>`J0Q Ǽġkd'boBϝY3cw7fa\%vdq*S%ټI.C &vI٥]H4uQ1歃O $Ř)՛]_ƫ"i] RU_b[1ML4Laji#6>Eg6R50ѭl.iz4V,SH02x3y 8Xk#T蛅\+1+uk ząϔ!.*\KI\SL(O eXTB9OWԊ"MH Ssdwt62ehձ*}w4=hϲ?ᗯFݖ+vzݳ+ Vr!~f [Y @(RNS4Rdd#[яwmH_QWCn.{qqܦX,EiEʗVk )|An4~!}onݍ/v꨸~oy]/Dcؼ[P߽]ݺV"v+Ϋ_@`ӺxX2rsN~˨7kȅOvrM&p)P-&v Ht"ǨzZwb'H ( ø#܀aLdKdrÌɬf-(gREl^aBNh oDdDFgd)-8 pi(eU,%eE.+ #*¨漀ǨFYmh2f-x7L%jQ9֜j I۶6M{̧ O0lfa}|P3ry|ݏVˎY_-`4HOr@LG@ly`.;Lޙl,?@C4#UVX'QޝI3VtH.+$Nʊ0xx|v ըd6SX4cJt|| +! 2QfI9ɫxk#+Tn9F9ñ5*zO4,z\7?Q!{I{ Uᇈ-rx!>\|H.wt0c!ߋ~R)|2q)eQ؈aSgŨHgB 86q{+۱,Iwo\6 p`33M;fHӣS5Ջ|\ypCJioX!L.G3ߠ{b pF;e |O+yfe]4Оq/w$UyUyC X/HUG]GΆeI X~8԰vPwJ^;onTk†LAHNu:$j/d艆&.Q2ձjΎvԻQb":cne]Oi艆jE4FxBeҋBS0ɦ x%YܵYft2g7Lv{6vnc6vnc?m|L%TK%TJ3'Ic9؎'|;v{ǮbU{}~9]!ԝpl 7v=h`7ic⿦(UU -AM]IBG:t[a^["%ϙ^Q׈'iy6zfM2[ELFn 7eX( {]|q6Fl"R21/$IN Kz-)m_չS y"m(⪻5^43&w6}EpR9XUd iz'aִ-w7w_2>0Ɛ!gH^T?cm i x2Wqss|pc-%F5ӃPWOaWOAF}.iIO5Cj=)Z2˧;g~2RՒQf"5%T*u+sK vɻvFkVx8@sB){voG1ұ.0cX#Q- 0 vcʪIߨ&)A4l礿10񲓛bFS0&HHNlGeTF6 /i F.DΚ;* |ՅF㌚к|ꇄT19)ĝ!\uK!0 VdY.ɋ\3m%̔XA%XB9sB\)S&{N RgtABHĚ,6rFUŶ,DVKes!>i,.l3A0@pfy L2ӽ 39-iFtw$)8Gh0S#6p-GZ"8RwG=}%UWpU k UBji?VĊbVPasdqo W{_5p4 [6D;/VC~q{޾]๳gABOغWTkx$A#ǨUa7O1pw ~}B+'.Q2wm(!j-щvrԙv? n)$䅋hLy'r T)}ϧX B yO8 +KeGWOG8W3(géz.+ $Gr??wy0d0^S/N4֘`M"5dwRZ9|%T~!ayܙ %<͜^^o+ꍉs/P?!.a+a2["3KA2ҫ;& &j$C r|dkg\2FHiT!{ƲX䗎 xabp]fafCI #p@4 qr!8ZjJEko aȁ1q^?J9cX;)%ϊ8tt!qc,s5:OȜX^)"|⪉&xt? +SHa%\iNz"e'ƥP+P;x㥄ΫX7LqQaww/\1Fm&,#a3Dʑaь iKmlIiN: ׫c&H<*+M> )&/'_JC2TʑXIGYΉ*%O> Um41YR`ȉ+ɐlxw|@vV%`oܓYaz\$kQd`"c(is.ʒHjwPK582Ңyrh0 {c l>+-DUQ2!ReޛL)`!0CJ\pD8yNsZ!rOBpC3e-fEK^fexQ s1h)1DMd6FM&ZS[X1%Ȭ(*#ue LK#cnSQ6L1IN V"H *@kHT%Յ- 4 kk1a"i1B@ >k4@|PiC. ,3 #+l8Nzz R<*+5* Jy3mT& v. MU8#?RKW\ f3_2Y^_6AkPKA6s*nf= ,+=C|Wayq\+~}&!o{Ky7O0ADmJ࿋?('JR?TP^__cp_ ] f3x6Bp%7⪥Ű"CKpR `=q@l<> )$S-6O8{& c-6*{_P9x |sJ2?eܘY{i2.,x]S&sۻ;}X\=刣bY`^?ﯳӽltv?_6%5X̊A /]̊~YݏUw?/~}~t[8mV'r Xody ީe5ǥJK{YQ - 6˭*r|o:-]eXgyp̅&4,SX arcɈ"тa2-M94CV9j9O~Ϸ|2sWpκA+Dsd^/lA Ƿ0FZ֐4# >0ƀ!U({"Æu׶^ Vxn[]mu- $muCޠrwlr= ui);E~ e Ijm7R}+r=#x R:5Uհ>0Ɛ!g" ByE8Dl9es%!>JiZ@?&=C{{5sBk/qvR/NmLd ݥnl{uF1/&3ѿ MPJfW6cd`;z,nD9ç5qZ=^u2%d0Sh Eg>+驛ۮi<ݶDӔNsB)QW$c=Ff((.:rwv8,NÉL,ƌ/bqQI6bQf YYB|)'V`0N1 3ZzN6 A+e r(-& AIRMl$HL73hʅlà;47G$ˁi-A'[5nX,-NK=ȈdR2  dPI \́M7IZbp I?dWDqM4FSNӤG?% `Of7.[4OCLfp*w^ey}UWux7Kma%4CK)fZR!,jQAgŏy[ /rf7^>>2)tG:ܡhT2_G :O7$+S!]JK_-a.%M* `V)JeVH't@3P(Hb^b0F-P6z9̞CG9!pBJҽM!BRg"3& A2 +=fŒd1o`hu"swƠEB5J@k)e0)ے 1M!5耘L+)%[8mjdPr)T B.Әx*Rƍ<&L]rUDZ馘DDH1wR|w7M$ D/g0”JY|n&IZO 7MRt}`Hv2/W,O Ij?,I}z|L7&?y侟Ϊ~s VzT+.`}984Tz䒬~Z7}0x3_8 Ii_G90Or+s$]pU7oSlS3 x]+ u%&1)@=ii݇`n snn%6Y:v*̱Cmz7 jFW%^@r \UR"˫3̯/7 0?^^fNFXI/swU+rς?&w\k߾No5{~}|?ˇ\/tNS"No|(|+] dm#n (6k'(:9Cg֛QpLXGfj!oMҥ)p.b#r0B G{s'EpTLWE"Z׃iWYџR( UUh,L1jG^ΕK Hy "lyƉ vWϺdSՑL/R OG )Tǎ'$'ٛET8[g?t'F'9N@*8'~e*@@sK&g3TR})Nx3ν^2Z[>wO\pNCHnlFB$9f 'lKCq]Bjmi^e{U^We[t-u a޳2E'Q<1Ng5!J++tA:]w}AI d[ OKlwIήמ 5!uYQk K!5-lCaBNuր&^K*H2#yt\ P\b!Halmbv6Aܽ3 ĸҰ{=mά)@lrirP؎KQKnra!#rES-Z[Ys4*#" ZKv Ѹ-1b4A}̆Tt}r=kduWŨE^ ~f}xY-O8qn0`)%myƉ)}].y)+B[qZ`j(-k#C~ma:ᳫ.#[S~ ul|yU T59_+^y=0dUp ^E!V9'hƛSM`*BS^{'𷊬Pr6(mBs1I)\2kH/Je]*J'D)X?m lZ[2 +0\!AͨJ21*) >w͔b涤 iW9ښ^kvZخf1rۚ%ruMxJU9pֈ@:j@)uu1+>] ]MAt:XE܈s?^TBd.cBUqT*(V4]++ct V@m[4?JzCEϷ0F)c1@@"x@IM& e10KVHw0/Kv3߂FR QC.eN)/=qi0QzPzYMOrϯ[%bEy:\$ 4Kj<NfC9uAy'\~'}Jea <&.J'O%h(!3(biJG6j~knw$K7hR=B|˛B@vol!Pse/|xG|ʿklMA nvV.ǞY?@ AD)/΋]og{$էUg+ *)Zs} }Z!biWMG"RTPF ݧ[1H4e7}ZYZ<ٶtY$xcF"짋N_K{ o?&(!#]ݭ?w7_-eoby^.M[Xc][Ϗ>NB ˆpww#;i>}MȔߙtw (p H nHŶG_bW%݇Ѻjty>]eO~MIˁ!AczҸ]WW CQ@MV6^<[B2#N$p#5("JwGdmAGPC"˛o 6/l0Ђh}\*\{Ew@iY[R(IwCt7Iiw-Q( Pky] [1.>tOZ{)-|QS[P@(2 FˤNRpm0֡;+eVKíAZN1pʣufsj\ ;Vxw?_`]\_yu֏t]i9`ֲsqs. hBz&z,p} ڵ:a)0r#XIq:V%g4,|O7u¦*tlh D:G1M<[K^2_߇Oa%uge:E=#lfke|~no)єĴbϙ])ڛ1ۿ&*-_پ2ې=|Q"<%|Uz7_^PU M5ɀ'wCHj6~whrTQk҃*z9pI16U"* LW2tPE( S.XA{juR ?bP:&uISZXSeSIrC{]ϯ7[i2~s Ŀ~kql̳b>XaF2[x4ԟأrp@7gÖ ޵6ncȗB ߏ..lwlH5qRI;]mɦDRV&nd{(fS;CNu,uN@(!@DT'$%rSPLjҨ Ҭn]0o.1\ɛNT9,wxr{'VnV*Cη%=c6J]]&zEcf#5??sT '.4cLj=)4T` ½D˪AR0uJ i!WsӽkSd\҂s|JAQ+޾Ut봳-ԧ*/v]A$t/t=`^M0_CŴjD]Kdy;ʢ{Vyw,'zuF3[ 8Ļ?>ۆ֢*#ڮ3 oW.' 3,C|1n뀖2p\@,/0,ʽMh#J1X^Tks1ýzw` _ï(hBT 9 )9쩴 w$_8Yf@GE&֘H(EZE$LCJR)5J 228$ 0_[.WZEh&i #A֢+W(7ЖՅ+א &N+QrYp 繢V* r c\(Z)t!EN4``;8&eMO922BSʜ)Zvn/ӗʡ )|ar e7F0<" 7!)uOF?E_ ;OcҌ=*)SwE:\3\nP0zIƝcOz2B$(Gթx4;'ڌ(HvT^Me d{傽Cɘ54eX@]B9ژ96H$@BpK+D> ސTSEZ3@(gEaR3h+3By!8MTBͱҞ@A55_CH2 iF̋"5:5iNH+i],I)DՒҙjElJ0)vdX޿. ͌;:S}sygٓ_+\e5=.)[9zy/;_ F8ZrzV)yeyݷI<sZl8y48r,ɟƗϚ ^&H7&+5(|1pY:h#x *cK7q- yK"{jӦsײ>1d~QvM畻\OW됂Sz2re=20ZctoB%AòB~~]Ѯ- S&&0@EO݉\uo( cZ9vEL# N^}{I!3>f3׏sPYL hκ/2O\'' vKIⱆ/(0J+ G鴢S2ѧl2Vxp-hLz,ʾYmsS(XƄJ44S<*5}k2jmmQ`bǫEñ;ȷ"F='w <՚]uMbі3Vj#/oxp%Dū1D$/ j^qŞX/d%ZR-|Q׷Kp C*b<_$.?z9ta~M&0/uP ܽ*QP%u-LThN(5PI!A2eeDT 1ef sZwAIeUub'1PcD{(F*7Õ i&`ʻѯFj](`TJrXRtכoL0u|VT[QyoE}UTFPcR<"SHqA`IyN <M9)(9 Ju&v>Ko6s Kvʱ!e`XU./h)ɐA-eN̛ ȉWYDbE'\#sH%m}$xsU b%,̍`9~NS` z`ȡ r$6JeB;r]ͺs kX\k[ę5Ov{݉&+ pɑ5'._!8"ZqaJcox?HD0{Ѵ9.oKQ2gەK`{8? *De=י ؏ЕvH;@ȐU&k8uZi$*Ϲ:i2$%iJƢI3j^$PD p X! >׵{ : 2'NU{뺘/!`ݮ&ٺxӣyy`Uɤn*iTFy~Ue?w3]m^d%' 1G?ը*y21]¤3S$/Y w S%6>&@ra|zpWHqٱKw΃aT3&&btA+pժQQt%S:1H gT3)Qkgi~ Z%( ѦUi;鞇{Us[QqY `E>9G5\^&91x2 6Kڂ77^r*w z5VRfI~* ݲ1 ̔q1.ƪ<ǎb]=${6- kQ 3 Z7rʦF9U (B\,#EFE12VJINYJ|0PccЈ).$ D"~rύ \dVBq*`3^3,D֥Vg%b$Z%] Lv?P56Sɍsǣz2#B.0lLԌǍBˆvv[;(gqD?_DHEUNrdOiwv(Z%SnSaTۤkVd P9JF+P@:3YݮPeU"8*C7[+Qs^6O!B85_dfrR w5=aG5ƒW4k Uٮu 8v|a(kN ݡ##zKL16DĬf+2F `4M2fS SFۥ)`Te't.Bv6CowԾ]QL"UZ٥To !U(fɞ!I@ '+@Y5 **xm;x=Pn*Ab8fG֣K !Y%vNEa-ҧ+ٌrT%f`%ŅqH6sĀx C{?/_# oxǐ!g /|y7KYŽ?AGy [`p׃Tyi;r^Lb !c?jvjpSz2ܠr ` R9ꯦ4޹YG1^nK/wOF6=;aH󖏟h?V=0cx;lægaӳ0tԽǛy43! wTٶ"s\'EF>OLp|E vs\] =\{8AG r.IWD9ZMW"'قLlk5S~s8Ps%C XYDC@.h *n^BS t$ ʸV^6}WoMA!bK`DlRm7k4<ګr%_҇IHk2ƓOQqeACiٹ-rvĔ›PzprBb*cN+7 p`c͌eqk#}wnm_}nQ7hdicLZN44(Eou?BOz2BEoscԞ۷wzޑe8ՙ?uZ*f R^uh7.(\FP֢Bq,*"?모 pPKgz >x5,Xz<{}yU b{ƣ ! "b;Z[@QX&JiQ 8ǐTJM L2#"̗fA%\Y/{ېP@ywmJ9V0{NhOJS-l_@`R8wdwmQd̿ {4y[@==/Z#`E:-y:?)j~w 0xo%A?,eޢo4ǟeUsIL.ŗJ@7H~ +˺RI?d ȪiL ,&t.^Q%ӱWaZ->_VTsj"Hα31^4g_pcˌrԳ Ơfrv=U 0Ω?Brm@sQ"MV+f;{;7/zIy[Q1{|=f6t RwuS]@5*9q-=5ȨV)FpY]qvzoKhmPl\|DljKBYܦ%)ǩ*FHJRdIF$fXKJ˰->;dVu76!! L~gH/93o?v¾6QwSٺ!9yMpDz%1lq7T,xպqh0y "E-PfH^T#G[= Oyt+%]o^4=XrI5,oFZiH`ekzS}EQr z%ՁQH] 405d\ I5+Ԛ}v1AR .FDЂmZڌpq`)H*Fp|,sTe6D=?gO˯-GfZt%L76)fnɣjh^ɻEǂJx0)w̉I?}]DMVǥuM,Τ89E&mIckmFE`wV6E@? :`; ̼L`QV@sA%|lIEn7ݖu.ůŪb*oGER!h`V C5~[U!Z~=VOJ VȕxcX$輱⤬\̕\%xü(o/S>Pt[ .oillBM@\߫ E85F2DVD8:Њ Efx$-EƠ=V;H1DhrAҍJtTҍ.J7tuSmtk+yكFY޲@sV\w r\n\;v$5[m:vSCFZitC]=rR*ecG`7cl:GR_]/Gl6C<V=sk_skWWd[[J2QY+5_gO/e~xPYBR{.W`ZrQ%#ߟ *ing%Jy~ ++kGBpm$S/S)lQբqBfAP1OhwvRɝ˛x RG !ޟ<:D4T/{;zB Ld8Afu:][ǹ..MYVL r#h_AGȮvkPTGɴdʹl%ѠR-J?5%& z[:q6b!/ ~eĝ0Bw ~ӿn󴲭uFBa ELU:F4TK=LznM1(﨣z$J_hBֆ|"zL )bfk*Q}7(M`fQn(z5C6k{c?J6jig,Q75[V -_@q,̟u~ʶRha>uYmVpfdE"Mx+x';Qi3:-U15d̩ƲIϘ\c*2O1c @ogWv,G<64Pcr*G2},ߟXPs%|ѯEvB?.cqmgg|xI߈R)hnpn05FxrYFWKl|Zv=["M+t@[*]jgk!W5Aqo;|L q@ rg?l~ob]A@"` V'4d[A#S[ =S1AUJb1G܁pW2ǥ {-B|"k7DU\l)}gF ѦG%4;eTSw L@"djVOT=9}(zPKN衃-Ge+ -?'wE%Vk:q:(/<{?Zq-LҎDm܍8P-š]ډ7~ri/R4{bt# l@̘7%++_`D s /?w}8*h8U Qm zWfL9*Kbi]=| i7gz@4/! 8 -)oLp'__/>+l :6UB^L69>N&Ѱnca \nn;J\XU KgWN(A:T|sl$fcNRUZ3RGh̻OY.Xߝ=4jY햔O/*߯3|%7.G"zW%>7x9C߅(*7`d*ùV##@'r}cǛq93?6J|oͧ]@Nim]3RbZ3-FS;_|.Ⱦyy@%#vw77<3n= fE"FڅPV?2Q(d YzLK~>&G*(M9y}Y!k:蒌E~i&|>k|*ot(]E骺(g۱IX'4,(+SJ)HK`AjB`5 Ay<x_?.Gr]eՃG''GM_t9ZWD#ঞO.+8x(h;k"0T{EEK9QY]jDvqU+Gǝ"%9 DEpE*:VQGӌHC\zmSޤ+QvvZ**{kyZz{1`c j YPy\ȒV<ûEa/F\KFA9(f><Ế^<#t4PʤZ<0]%q'/fѵ3H.2Z곋c,Y.f 㐃ѭ|{`>,eydnhWv!wxEe2T{xTPj|a"u"2G1C<>"ajVB__l[03Qf*Cc\OZat<-k+}1Ss|l-IKD=;U3—G^LCs.p$% )5O"=0xqku%nǣEؖk0hN sxEu<c_@ ^t%֍(kuPoI7{ߔ|?MlqYرzȸ~:A*]%H(ղ`3o$ztE85U.0 2HKG/Ryʉ0qᦓhE<U´K,w5U[y_dSU)8\@nv\ F&'8xV8s&]U}M ]h@h9WςDQnp`Oi)[2 [z,XWx)&4+NJ9{{M5Ks9Ez& y5lV(YW{df|4n>s"EkkXu9+c4(z.ʦl>3uu3%թƅ76$iCgJDėNI[3vRۚ'{=t9ԉ-ۢ={*oQlY +^vu3JHծO*>xJ3qgiEsXiJ HZ6$8x#TjgފapN 'N!7cR醲*G~6 ڡE9y]V 8ӕK-X%b.- Ʀ(]-w/Qbo-5c 8\fZ, RBaCb3,ǰsKM }e"fm~;D?Mzܾ2:~ZWEƻʇ&=Ȣbh{C!8L4\*BK3Ҝyy NHٸN'{Mvg"%gg/ g~u:7YD;Ų vwϲr~^XdN;2>; 贋Ww{cMY|lM=xshegq8)gU-A}ҌG톣R<5WYF.u:JNg'׼;!F%Ǯ]i>HBO1F%{d'O%d6.L)]mՑ="n"贈s&+7y|;'r#(m)pe &G+W6ӏ%pX=j&)f`]|\;H*+%-V#NؒGt@"h \Sq˂dv<%+{Q$!(!cG-e4| I{i*ɠQ@_~D=-iPUz`&ƅA|p {lIɹe fS +VCekϟsa1JrK("tB2YQ1* ֢ ^>8Cӳŋ<+N]%7Ӕ}vfDU繠ZK}:40JBQApϟVDPbAp&\%6qfğRGR x)P%7DKYZKhx$l'%aseܒ@,2JAwM]|fî1 xFL eDuGYGGd[ogR`$kt:RC ) I1[Q |ć2P͛6pٻP"vкF-c8Ǵo,֢! heǍ x@p)*>k Jfc&ݪB"D4bzm>K΃ %tVEF,DF^V$6Z+ Ӈ8Gͮ.@xtQ &ESD28@.J]ěYs^ ڡXm_Op9 ׏ ͷB?_>\Q;rM#n00=:k 7}. _BsS~Lrw(E8T'Xq"tu_$^Pks#E(B|.FvYㆣA;P-OR%i!F)oS<pߠoc1.Ζ Z)ɼȨ@Õ?,$WdN:J ⼷K%?>9<'Qhzk"1tBb.2*cpzk쪵! ӥJ4Bf M'JxR*h MY=s{]fꍯM;ӢF-hikGj!εW6(~еMZA-?Y `3z՘DݏZ16)"rN(p~ߒ;!U#ͫ`Myv'+!Ѳ+w̾Uv-x4+Ռ7so[sw!mK(5!8A=r*䌽<ͪ)N8w?7|Q4Va$-N^:ƃ]!?{8cC3ഹT薲9A+ rNtqSs%Kڅ۬ud收Z<N#?MY'A=pr7񀌶? M91 "="hB? T>NRrT䣭B5k;9岡q\ж Q¢I.d3Zp}{ѥf; k1EL |34Eo[(dXu@AL&w.}ORUQW6 |cr"#a_2xO,L89飗i$p|w7}>~`)>'M=[W,\O'v84 wgߟT,MvB2NTBQ#q@nPJ+)JyX) "(f)VO{:).eR lhXMuVQk%f"HQ?۔ M >~l)wIyp? ]xa: eWSğY+|G |G<&qR0RrhϙK% ΐB:i+_3R-Ja%rSuG]t903Dcvc5]2?s0o0z K$ aR3${~ő0\]_Q胡(ڲKNDSqI 2QYk| 牲R@΍ (A;)+Cl $A:!uR_6R jXBG7FXL "h&Qq]#*P :MX w7Wg\]r{+kU4&<!* {yAeCf e(TCS9/{λ)`9w1 SaWR`&aÉIOgNa9|MQ U܎)v"E2]G嚧9fܜߤf<' .ɺ=U>4 ,~ǯ L>K~2́xj'XeCR_77CPْ k>W?tZ%P?'G[Yۃ+J%I* K]>k ҶY!kJBc&ny[l(O~ > EeͰЧ0Ga~7MwnToUZ`)|Q  aB(+O$ÂI 3`rIF/?[~e1Y3O^*E`ޟFӻͦ7 FkEJadƋC/^L/d#x2]&ݤtwSM׫|n*C_ ~oK|b8ĤZ6=|d0d]no:+N\2."_.ʶz> iьWLɱ+a0% Y"-5G`zhA EL,XcX Ny(HGPִ pq3>q2}Zs!|$ӓ&turTMB\#M0{s~xb3OP#8s:Ϩ|50 cin]OՇ!҂ "ҧVf  ~DMzc2٦WzTʧ}19('4mq 6w:%R(a1*4ZIuhkdLt:ı?P-svPIZHzf.UOrՂl24un͆/Knh ~~ +zբSYt :vF'q]n>ג{dMtaY]HWx7pwo-:jh 4r0[h #n1T?,2i ^3!8XDaE*MX;uȼ-U䪊cJG Xv&Q0|l7+eA8%HۭjKb8̰ s |RK# ʘRJܛB_CDdq fZ9Zby T~4J`#GH 12Qt 1gl8:`0_<دʵ8o(o㤧6,Mi޵~iW!c*@p8ͥ@z6`2~Z7hRNN4O\Gw?E 8o6CƢ[/*60́Z,|]%A{a66LfTeJ)8=U~_|$Şʁ%VO.ª̭=Xc1L1pddYsSFr#D7fGqp#**~P`SHqZ*kٔP^yus>ͩj[cVR%YP|ZoxN*`k [PocٗZ>πj; fX?oky{aq1OA=̦M}O1 QֿK5K}DnxڑgeIjŇ˄z2C E>IbD)nǜ;־SvVPG"(H{0](4Rж0cy`hтgy?Z.JV+>?Oocriw1&mc]~G-Vng7ɳ|9-ki;-n;Uz6N.e(@ʼ~K\)@J!0R 4/LyɅ:H3G(n%IeWA%̷5M0s0!24`8 ?=?y2o޳e\{+Cz ap,,B .( L<ꕂ2nzJSm-LE .RD W|:[ؾHLlS'{Xrϒ9B.ș+p`YDV Aki"M݉48%WV+!ztX6ٿfUt`Quu=>"%,r8dwrEzb@AY$'R|52@rMD _?FJb+8_6X(r5$8纹c |8X%Uo[cSjtQq;-Lq@5.ywj`KEXw;;g3Oi*;bީSew%%A>-ݔL3嶇YG~-6.7< G/W)@k'cdž]DqsZEhO(sU8 lO3خUp$DbͽBkb\Ҳ1L" 1JivAeařfϵ2"v1Š[AITtT^` %)ZD N. cf:x?;Mڸ!`:,?XdƦ$ƠJAs>1*"b;L>hʫrB!ZlWs : 2:x1,C?`Kmrl$Vy `aqZl|6")3=u";q+L#5O`耵S!,C*(.x$t usA;a)#PRБJc M@Ձ[ 8CY mer QsD'Z=n`N_y%4]4ND5[w $d,ciQ3=i3V# Ӏw͉v>\\2 )B)x4fT-1R 7۝i@vl|C^!k{5(pU#"}bHgo@._}gsQIy(Q"-0TktCGb2D&{,ow(:nCCxNZ^' %=z3-&*AZLm<@pF;W=_DS2+~R̈p-' %9rlVcr|(N譲Ƴ|rٱS_v^ ;N:S}2.Wwmr/>| Yg&1ᴢ_0lK}XS*&3;ϼQ ;GK5[*7•dώW'M \mzl\)[ЎN@˶@;*b!SZBOh'<%MįH)/^쫂#voXWuFspX UEn?E ɫ8n)T;&b@|l | IlbMz;%Z|j9H^)ݦZb^ͧ }ٵk42ŠHwUMV^P GY=ɍe7>r -da栩}91 =\˞ɴ`YY6nHB"%S4o.1^8jQ:",T? U#8%P󩖈j|FX:  kyHe?uzid94iɀ#^eg=Ɓw<`lP@ )+fH-AP՗wOBc|{ ̎+Ht)IG/6K.~j1_Hk$Sv z*g)ͩymVz  G)G.EWZv|l%nuC^/Uc;`OJY[kZxG!ǒ2 O:b{*s0oao3 ԗ1md}IHC Oyo:}8>6'gXרFoj Uu[6p5nВQL$/8B Rmq?N>{^) ^MQ &TbjԚOӞ2M9.wByR "Ӧ=w][:/Jj}[ǖcY1cM̴NJmD3&ƒtTBxH :TEI0DtchjQ3 1`ғ&\DB,Ҽ|wxۤ:8n ƤR>-ť,]? xGy3kHD2uGO~9j( i3n5eI2FK,2s(@ N!hI MІxZF}).E;U ;|Уr>ηrRx4oo΍BY9+<rX t\-ܘ׎VbIl:zVf޾X`%lszVw8n>Jn|(N"+D2vl}uH,j䊩}h5AuZvz92Kl2=9xAiHECƘXge/*JZ 1 cׇZ渔}"RP:w;*&)Ms"8* F6xࠑJ8QbVVێծoZ{~V̡tTm}G\V[9Yi8CH7.-2e\n!9hV jD}zij!$уe\H5]>& itx眃;¶!鲸,w+w"xW-Iɻ$!LUq 5> *~|l=|UbSWo\u \sd'%!*oc\JQ"dtBkZ/ge$?__=?JH-J^z!=X!s`1X1V?1Yt.l_W % zI]̮dqLN$i ^̔v[϶>5VWut{׉B #x;R&Wͷ}p;[dB{Zw.|"d͂ړM{Sz}:q<+Y"YoQ@i,&cDBe!5ge>H`*RNx//zqvS"y֭؋򃅖4WjnhL C&oXOjW%;f SZ[m JZW"Ĥd+ L샘\b6|$)ၧ,; BJ Z ahVCjzVtbH#Tx2d!P*IdIy .64'Uo X\;bbaWLyzž6Z$6k(r-!Y,mw$cv_xȈv\H&d$sQ)8X.F NKڸ—g Zd pk}LjӞ0q$|dQx]JK5 Ü8l;'Zc>~%tԻBZi3i-%CS{N&hLtAq$NLuMT 8g lg=?vcL9;eZ>4F,vkJ3:y*u7TzA.$ imsT1_Z[4-VB%xZ#K_^Uku6mXZIzG1X{iW}7g׷X\v1T^)2Ѷ 1#JB ͋Mɴ%`&#b45wC`*?_ջ`o'gFlVl|]Ⱦ<}`twn)|$zgA>px,/@ [@ŒPpXcE&pDypy4q:A4\:kd0- Yq4E/FѡՍ <,Wya7P )EtCi  fSN(TiD 8TR-B1K|82c]<|ە\Ҡ2bsgʙ`'*%$n(%ٴRr<")7 UT+ J;bc+5dxm t!$Bg}oj7(UъA贾v;IF<0vkw$+hLN)q? _+?+'ȧ=iOp%(%8r+i(bd!鳠ۑ;?ϣpG\M'8Sv/$ZJ'yFJj]C:w/}פtSdt ;LjPta:L0s>54 :G^H0s&8v&wqyq2ݛ+>7Uֺ(1-: -.g-grsU_Z6SդO8hdm渄E8kdxYFja hS=h~<)_>ЅU篗Aث /|Y)ɸVqEMt_B kmwSHF n, 6h|(Y./t{ȟnVæaS-I~R,0:p=Mt+VК'/B^+z.Vd 'ngmz:OVaqS^)3*EN#Z KDf+G{-B܋z,>n~fÓpߣkC1@UsdomQ̓^(:wG+4sM}~_Ob:dSU}2( .1"Ym6ZiAuvj#*asaQ*>Kݍ,5OU~JHp2#B&褔-wQ稴^alK>KLEYx@/ؐӄzecty@$Q0 8-ZpG].A|zAoj.vhPlˮP}gUgc_SUҩpӉm)ӳ10}vV-df.x=*G of"pUF:Vd9t5X%+vب^e0l^P_BWdla$l @K"5_uU3:U ίJ~hf"X-2Ö"?\T"_\gVC\[|EP{pԒ>PP|:FUd)TbAȧd{pKa~ӓX=+NG 7XCb[FCAQ3[߼ewyFF2'@R Nyp9eJV'ҁ3p%=$&栬t!ZtQb㨨hh^SOZFB2>Ja.}?L 22@JZ8|i~:): -~#~^w_=k_ ݜ:=)3d^&Zt H[r0JA Hk"jU%'qZ8‘p}.G:`ukX0]SdFUy0y՟tTqf;N9XmK`xy]sA/0+tln"eM RNs<=]*`ghعM|KZзx̉oAͲȹO}A3Yf36!j|J2l.׋e qq`yP)z7럯o>}g!5 g/0ZRȜ26&) ?|"]ޜMzY׭j.ߑ؂f}21L@OU>Үv@09ܥ^Ef n1`,NrǺ5M%3cD6=Z\oKxv]9ܧ5{WkS۵nfHo/X5sȲEY\'t ,o5LBJG[wrz,_*Yjqd|z kBNr:iYsEn洬*f0csIdڰy}̻`XIÔ1w}d]y-\=Sh4pL`Č(afzoSm;F j?CG1{$my®b=&oSvKدa*Ti7F?m[$ev =qx4`PRsT/*u`$ܸ) (XRR%  d#6`H(!B V{H'ζ|=Y4Ze9w1 y'dd9;AsltJ*$ޒis]X> ozo2g}\jqp?3" k\L-3a]Kaua略!Oy4)?\|<'[ٛhV<K/G7G:7Mǽx{JՓYq:;Mk.юb>.:t&fG]z' ĦFN+' mZI(_ f1΅ bŌhyc1%p$11!J :!rX1I _MKg{T@Ԕҩ{s}]LLTD2-.hgz )*lj j,X$ O6%FiI͠VKB>rq>]]R_mH%ev5<] Ua,(CxghO>}DhQ,"㳖 dRk-6BET1y]b@:]m'@^EE]aM${ |шR|Z%*4q7몉pepUյk5 uXs_YGf3̦z%jq^1Y[x潟nǭTp$۫Ŷ=f4W5ZP uL3NcoOAu3_6n}|a}wiv'>R8X2!s/k$Vmv;%Ă@露&ueϞ>)r2g=7<@Byl\1%pAR#pN!mћZ4d1m)bIf噶27p|3ryy6΀FQRxǔ w1k-PU|#8[!sss m,nڈreP4s_gꨗjz҇xN%qbn*51 bNY% u^P*JMu³CŸ%.Je`%ZZVLU+f_\EӠCӏqәA̪_uTQ^Gz]V;3X%92ŵ3l eT)'_a'IG?y;|-͝]Mt9z;QEX͟fgfD~Zai ܤi>/x"W7A#-T+EW31_MLUR2j Y0М:/2rN%&D&L9Lh3S.1eeV%19]}: p&6L^^,*Be1] 59ˢ`aU\ BMl[xCc2%.=meSC\NX)jUo+Wˣ@Tz[>^aj# 0r1ek]к@q=z@%I %!*˃=1N__cA^ee@&j=ܩه~F2d0+r JnIX6P$q9J_&P<I)O@xH$ƚXq qmP/ka"l8(;ERfl=bx=ݔqhKmݴZ ETڿp$#!+;t>KxaV$\}3D5n<>43 g r;4yFd%ꈬ"-f!}XJ̕*X3=ԓlrѲgںi3⪞+Osz% )隰6tx%yۯb VL㐦কyމ"CϓbâF>:jm.*rimo (\_%VqamH7yBa9kz%wNݺ}3T9f.4SCv;|wMV}?w%1)FecLNqN%>igh8Y) @[ w«BR]=B|)$GPyY|påHj6wWV:^FU0q!Jh57z:}wJ LSLi=ӯiۇfl!!~'xWk}1n=n$Ubo8 qZkb7%#=RV^eg U(O3&FQX[P0Eޙ0 TiI0OS 6]*Y/գic#KlQ@5ZkG^.5e@m%?n11|sY/K,n0/LgAyo.blgDOw>ߖ=w`}x&D}TR-}7 vN5r~g7i7% n=2u#n!VyVxI/?ݠZb$m״Sh7AgG,7f'XNF4w 1: )y0h< .3 mj:nX|ڰ9 pS HI+l>APGFTr/:6d?\tf"Œ@؃SfrGBfaz Bgh)7߻no2$Njm}DM3N(ɗP)IQ8;4~RN$%N$&?Iƍ'+紨'Tܝ9h;j#E-tp6߻p,ejd(9E/igǺ 6V2+RH̋Yߛj(z~>V $%Hbk]x|QPG}D6@*5;ZvkͣUEׄЬ$7K49a(%gVLfs,wPZ ũ"ưQ 45mk XiWs1wQyF5~_N.J]69d0W(;Q#+egjiSQiot y?O]\[>%a]| Ȕ0ȍv ޹[0D஻$MBx}I:DP-)ib:pEcEQgqʋª9~)T~ hp4. @k0*wݛ2JV&..hKo-e3q"~*vOq&\xxaxZ\^IYZ>($ 7ͫ׺&{ N=nt v&16OOM1)b;'+ 0xu @[U=lzՄ#1*J7[$ FaP pg̡ tZA='^Xst:ή弫3ٵyz&پ3 ~ijte[HQPKXH(ʵe[rGʽl^){(/*;//TCL. 1W|ݛ[SbS bƔ+ym_U8ϊ$v)K量Wjv0o\{Z{ %ѩaG g^?#tzPf%~Hc鷮aUYW+BOڋBD[RG^YVbd#$,u;V{nk^Z&A$*/>^m./IMsljc;g_2[ݒZ!cK& @? a||R"7sP:5Xoo%?_=l.+tEm)gpys/C_UL^KxhO{Kd rh`|5FK٠`IOlfobz_D{n2I0g$'ZS~ZNAlށ<ML hU$אRv.i 2y:Yr*ytzь\A3L)*0kQ_(Wfs10S 1agѶL\N ]q,bC+D/ { sɏpVF.Ԝ O>E'u񆂆_OJ2ħ4C>C>t]CJ<ɒd4xd̤ʂ~Jz#x'\ċ|k]Qهֈ]H&ɑ|PԎF1/>e(8x% gDtsۗWcfUrVV3x)"D{ wO S<ۜr6Gc jjY~<?H ϋǟwE.4g]kgJWx Eb"DGo[}?_dsq|BIڝ왃߬}u_j̇۱/{Rjr@;lj'i6݊}[yk}jgU85еA%[k{u6w@BNA(^ |Jm#Wb~^]f8hJFj%͒֗\Yr!Z?s}0c%,itF-A`5Ca}|ƴ' wKiײ9'; AU05?xm۩:)Kft3*f h1pc*F 4º3G;3]ݙXױ}zQ \gUWulc?M{"XbP8c5W~z\{(ud!d=}{C{IoOڋhw }86Xt-9qVr8i ~z}#gou34#F7c׼n=nb7vcqƘ2AiU1|+-I+A-ـ9E&KfV]IɁi.0'Ye1U <1B11~]`8ƯzNX_:,dOp0w=81ፄ7y81 o,b 7>qQ}QMq!Mx7SN{{g"lhiJ[" K%57e/IlL~BB( 8gH/hΏ9^xxپ'J{eGgPH=Oi:}:+-Ws4VMtBdB`Xj=G#J"FUu:˜Or% Мnz߁&u3spUT|7I >(N"fr e\4>^Ȣ`έ_~Ats}.l6=,;hZc =/[ CO D?իם? 㶓H((;(`S2J11`~\cu.z4{r7d#7v뷚"\^C~*HB)@U ]`)/OFS$5\)\y1Ш"ꆳP^Z@6u᪲9흋N1wNU®ZqwNUy{A N!PU;RϦF^,`װ/upy'V(81rDv|㛺5XP䫅uJ)`ys;h_u~M_ (5)> o5郋u=#j>o.mzR 䔤mVqɜ7Jfr WFW.YD6S;;ߓZ +Cɭa5b H BEWBE ~nC1LX(+T1NfڙFH:6e(Rt"p6՟|@ŨkKG0dvjĹ}ߚ>_>z{㈬F`Uj$[ ȗRo6:EMP(u]Cr%Cm' r{ Z})nP# ^uNq&r6}h| ge/Psg-3JKNJi㴫xv-E󻐋Z\J={v*YۄvoCz)eo"|((‡xCxkbdl\V+Q,L 0!8a6Z./=b~1_W)⋡(OIw>=`wOOSgu<Oa=մ$|҆/Cz WA[$ڮ.`os.i߳*,1+EڅKH>xX2 p,qJFA.DrhRdNEuF؎rzɸBЏQ(G%%0dZ' )(&?5Ὠ$';Z%Ub&%fDg1)ű|^0uTi-`c`XKRg;vvLV+`Z*+}–jmF{SRUֻ&''lVOOXK%;iqy^w>=5qHC Ú[޾n33}f1`q>݈]{E\D801%WLJ&YxxD,+/됆8O -J%WG 8ڬda&A#!G4o6%7;1A`EVe sY6;FR,s͆Wg` %ۇIJnuQiˢV)!ȬY@43R+fL}"nᏬ$"Q_ TLNŌXei I9bi& (T6 5\bi^]3ŵ'y(m073 C-^dvF-Q0U>1ţf D%ʐ/$Zzjɂ>mgȟdu/ ͸BE4+O^&_T ^~!{ kXѭV2g}# lꧢ ~'3H`(;rO/9LZ9cZs>:PAKX2i]ݖtZ-W+TNJ9N 4__4z܏{rՇW ~zE$-aϷ,VۛʭOw۟,(NJ>1 R#V p, m߼.obƒ˚7\.^R/^>k@GuTY)XK%dI$]b5 B.S])2vVLOOi }ʃ}d &in𳖞~)ێv16\vzZ*JLCKUDKBy>8e-yM I>-N ƞwRR)]L殮jrE! -KѢxd'+aACB栅@}4D3َTDDQK"-yi`2C6ht4Rcf 1f$`jJ"M"q{!uxɑAe0VA1wmm8:uH4ۻ/SSoSk8I/(mٲlJٝI\$BP PUB*|>BMd~?էZcymUʪ2JCjW>Nod"Ivgh,[i6`69ABp ӛ=#<,Du`DC>|46[}z {[n|>'T1&9/1\1@L[rk*@\֑Q.@-:iT9m@ves|GD:6 Iא`@^],gaˁr Oߔ37@-r cr ;I9d٥Nz߃Ǚ|ػI c ]Ȭ݄y3H]km̈́0&6֢j)ԙ#%.4=bRL I4OO \H *ͻ'.>H?'rzoS9,s`FoݸͥóY9i4^jxV>A׭1 ww+ޫ%k*Wvp]U 2Zh™aNÎum5h= ;vX'\Ē%:}3%`C;e 7$ {Qo2ޯMo0\%%#-&ZֵѴ <_8c?YN3]PRr(_~17`sgἴwQgp܏)]sezd3+p8ge:߫7!19าH?j՞9 Ej1[6XYVHZHgL}2Z=%ۓ+Cp' ;R>"kt:ν\r=y-L&Q\1D̵Uב jm @(kdQx*?D:Qce=[k-uь6,0:@,Fʒn; #PBVVB(^S$B4eɻe=cHY)9A-&]*',"+㝰 -聁l1׬!xrK8땂׆WW)_ݍDخ[8|u 6!eo܎NA߾n+OX 0[<) Q7Gr߯N?qV-O}|ަ~stxX>*b0}?eFcp]qc8:&AaN~󖻏 3hyDF }YlI!$ b%Zs+/WL^ӪV'[D3E;HJ'k&JB;Vkls,{&}jy*DXNX0jB}GJ+8:1ɳ*Λ(R !](HdK _yX>D&EdSCTdjU %@a =%a\$lh~/ #P{"'[K2TS62S(;"M[<٘*J< 5wf$ț B>:KushΒRxt*0V.TkU}Hm4DY ,%(3 ;Fز9`:k},+xkW~ Ol9NXkԛjt.2h𹨁:M~fGg-CeBIK7mpl~t.D|68 [.Ѓ.5g9yBƝ%hUlL^oеQ .! C$p̮v&;m~kӑ" (+M;H >B$yV7% q]EmC9w( 1"'!Bg&Xس< PA$Zkj%(ѽ(3£*$Da'x6=Sye.!&e) < @O]6M:8Y Xug V0Ƨէg-W='+/k?M (NY&,2?6tZ`׃H!ѓ8"EgB$;=Iל *SFY/1F(Eqpk&Bn^EA`&p"NP%fʱlJM#P{y$FR\#LHٚFFl8.hS뒶?jI{FՔ ϳ(mM쨤=U-Ƙht/?5|L*aLڦZɂӪZ\ Zp?jFG*Xd&тDQF x8-Z%|[&Zx-ЂQ k {R %bK rgSj10eÆNX`=bm6 mvEE˳*1)#CS'h& DQZe0{8Q[jjɦ"+6rZV+e^۴LF^,1n^S'j]_/.?vC|Y.bJh{1>U!!|0zGOb*zu˟vhgUjӳrjyq|_oNˋ޾+?yI g˓IBpvLGX9/i?NeW֤gu;|ʿdo֦{k}E^k{sˀY[XhVd(XGv)o=yt(eXI!D. \}+֯Wc÷K?3D/f1wo+^iI58k9jS-aފM(p^M]~e`[Gz \!}l$_!̈t{agn <-N˔l,7$",VuX2`T)++3"X &K/gI3+`SNXf;nZ[ݎ[ˏTƣxd60eo97)_24 aXY~o#)S`vhY,ɒF$?$_]uqm, Om+ 5L/?~/y-@TLC=ڀ[ܼ]j.p`w@1gY.pՠƸ5ydFx } ތ!zdv V5 n, `#*V莼5k{~EJ.?FyUTy"W7>T@לT׬|(]/²D2tBit__@)}/臟~%]_O>N I C‹ =euBMӉQb0F,q"GN:oA޳vA~Ca'vQ03z#{ڽx[}?sjmP}?ܥԬj4 ^f2GN.ܔ<_Elx5|.WK ;up?%N}kվ1=0.|dy&{-=1$qs@T| $;>&64cs L_w~D~-=Xݪ`}4ڰGx r2owP!iD荩OTetL}S2L];L+7?(LW/' R i+Vi mrJ1^ 2^ttse -l`=16 dݴ>KŌdVBqwB EI)pb&J4cSSaL\9IdhR6 AsˢQPZjF$dBPF*oԈb)6r0p2CMn>tOܥỶ^`{/Tp.GVN5Fo}ҤRem7Eu&xbk9%t,tҨ(db]cri[(ep8c}1D%=:"=xeL!Q*Vde` B ItZ)pCl"OLf2U0G"y*L`j,x, fe2 ePKaA :%r AMJI)##,1IΝ2V@JX)Jl[FfbiH$ORSq݄No`Qq㠵vC|=9:}&TE!+k,9%mihy]X*޳gEBvDv'..? 1cTt5YJXhd,*jr2랙k`! uƃj{7Җ7:7of~h7=UI{ל~hAf7;ъ'R0H} 颲X1G>s:xVcєqX)ZǃE`,J,xnM)>Bf0nk4VJ)րl/㠔C%9;.[y㣕.(xe2# 2wYKt@r -ڑQHUYߙ!P+3@ MRԖ5ƨd9gwzvݚRKڨ[znѬbўwvI@RU`J?w|rߣ)<Ǫ {3)myc@>6w>YnOR1CTR+h%4{ٹrӌkG/<\(#{/0Ljm-YDX^ M6(\hlQ,B:j1G5Sz"Xk.!"L8ybОwf!C!Ul־\YƬ/âOӿ4H7eێ3sKl~5gmc.߄NPÁS6oATg0fsan4r{͋5;e:d ciHzaRE\CA#8dD%>ue\=zq@ۿÑNe. WŽ36Zh ҂bB 73one{(Cַj0\6Acx|:e%&,2$X zi=H͓>KؿFg;EgzDA;Yo#mo;G/h;|K:*g  TOvM8hAj[SUf> Q'{DH^q p+5Ձ!04ܫ7IX7X[ ZftG(A҃T"eGf]T&Cq0>ٌ\_;Ss[*1 9XglwզBh DY$N\ELkn 䖑@1}E9!e^>ڔܖBXP SV!3Ev0XMePVfp)y靁Ljt" &``$0XfLvSw!z/\^Dd!ZZL gQK_b aw51(I^rz>XE^r}/R̡FnfX1_K)SAXP%`$XIA7c"BC[U/t,ȻsQTl7R{$OeO;+ |{ڻ%&&XlM p5KU2G]NFSGwO*џ'DYvs.,.oW7w$_OSNR?3ښ.=9ʼnhŶ0`1us]B?7wl?7'rO "uiw'!;I-O# L~Y$i!nёY%jښ^s? Ċ3'`/*E)GP % @+$ߝɭ LHOx+Z8iHb PYC|@3'Z?/kB`E8M D\24u`QS I~O ,ZW3 dZze>N2%#is5'~lx9QXQy$ədLE{2VlLG$+=µJF+EҐ5emJ}sR0MnmDI@*z[܃qOdBB .wh=kRLwۑ@I yñBsw$/tFmBb.[‶% 2&v%[ZCi4Hߒ]ή[=Hڰiq8vL1S(C2)g7;իOrZ݇#\Ie_6Vǜ,+Zr^ G˙9}XOEa]8-4'Ea36G+ [`QZ+A(c_y²# O(!,k s= #"8f>M][X-ѐuG=8-lJ;bP^|crҕɘ~?6v O\ ^@vW fZ;EP'Ad>lߴvfr@@>reyJ 6F8!bf\Mmd9l\z=H 㲆""CšhM"m(4OhwA tџx[TuLegBzY#5+Y`L(kG6;;;;~³FZmUu!Ǎ#I_3EyЃaifxYB2nd}3Ivx5UE6,Ū"#3@jy;#?-Kq*ǜw5Լ'_U:Z=.. 6497S1[R\ķws?]>K`֐49S-wwGߍލnꯃ>r?]-]Ն79{Gq0CA+"cX .Ve 5-b}R;џ~.TZǕӋN KF_ oPNi؂TR:D*i^/$=&YVpSjP=jaJhaJe*Gq A;*(DyMwCSZto4j )Zs/QH XiAaO@z JMHQԼ8 }NJ{cx95Zg#Q  :uVx)g$`M%^(IL5eN$']қQ+ZK"2K2ujVz׃ՎyN_]M'vY%Eq{5o)>afnk*~.AJ: 2^JJyeG1F/fITڇu-Z0l#X |^5a*Po܃!{І|&ZdSJ^̷A-A>w;_40e1[ufz6, 7"uwN/» t Ȼq!@Jn-Xn[6EpL{#eVn"Jt޸xdrA 'wơN& ȧ;;/ )AvjsS!<ƭ͹!nmвJjo*0T:\"deC1h89w8$|fh!"J4n<;f`TP:xٸRkV$k٬@`H0"W,*\$ k!8p%̪JIA+*Wv^tq#GJ)Bj.\T\| T( MaSrc|A$BK>xjwIɍT=|o>o_Q|.n-bO,tսm JF޿B}(vEO!>ypՐ@D @^à.s42!ԀBDBĐ#׬V"mұ>;9΀2Jl%3^5k#2~"x]2jR+ti+Ċ % EXcȋ;cȹá3|DB9I=߮8փ8Nl(#{qgZ']t"r 7qMՍr{?Fq|U<~vo{[#%-b5I6,<& KGv+D" YE 1B1L >?Dq՟pWZGnjo8>%o`~% *G* S*Z9CmɥXPVף'4KǖD h[S;EFP>Vbƽ~-X)dP#LI#Ae(Tj #T\ˇ~:mu2BUpRM4DY5uC NcK{(xɎ}_8IUBfU4q*~\=.m 0zd0z~qc_i2~1m1 V%s1V|bwlzqp&@Rk$'ZDBҳD6R$BZQG*$-@Ұsg%=Nɽ#^K@q9QWK΁w mmi#H]i#rvJ9vÿ~#CAf B 9{biνT!x%BEK`Po:!=gɈ\2:fqx{&3<K8S0úU.D%pDTa3.>XfժyXXy%>z*$ gS}Av)սaRJ)<q|c-l3BYA h5zX:S | $Ous x.8iM&QEӺK_zGm^(ݖp,S86D*.U{[ Cv!-=m& 'LXŀJK&*XqfsU%yc9(܉-94NJP92͉N3/9@bSTBcUѨM%HtAAFƙGwQ .  F-;uԞ  !z,x:9fEv3=RIs悎^Ou?Jr t* hSӏ>uOa|3.垥ϸ\E%ot%fo.^]s7ihi ,[i?a[?F8Ipzd8wzn\S7Ú(7ԀB0u%DMgQOvuL4<#3kF7Xg&g( 15E걋mbJwN!S0Tյ2uND :eJ-GaL`7^Ri *MgQ &a#&oF6u]u:,:C,SVWrbiUV e}o !WxC޼]%$°l/~:" Ѫ? m~g&(_{p/4ⷮc=m|`S?Xռs? p@jpBw77^IS!=c,0_ͻ)_wM+A4I5DVSGxs^W֏xguv2De&K14$S$P5(bB%}QI KO͹Á$P9w0HjU]K岼&رLkPLlN9w2?wIk-.o" Njo'o!|X{!Ť)rzUzK;HO* T&;O;]٩r˫Wy:ya*O'T0ZJFT Q\1bS V3ד*l5RZkU|8u*]5AWqF$5˅j%{BaS z\Mgax)m7T,WfKW"3[C+5Nv5wfˡZ=96oL^7r+ 9!Yd4vKnpEp"_`PqWۓb#i&[Z$H2EŪQ[qhypVoOd&X51QkKOӑPrbO d0"`URq~Ȗ3yn/ n`VHtYp RV"@Yr41blHc>MlX굇N,JQoFjqgpRo|[UD#+g-E cT%h1Qڽc՘FNQ[tv/r3=M$ d#gxp5s\^ -e\6o㣒Fރ.{We| EK2iס;3|j?)-*XU`WmK^n RLsH(Zej hKe H3.kU.ncF'䟗OC}]md&]?t9 ŀ,4ke|sZfyrSu}D rM%bQYu1'6SfjYY,? 0UGbiSN*L&ծ-U 4d*J%;NPdl@F;(2'ȚM w oڛX7o`+ŢȂbP~]Pܫ5&VHwK% b#,Ke)g]zC3澞Ɵ ;(_yD%_~fH( {\LB箏I.5? Vcš-` 7$=7R#ZBx٫#aAB+%HfA0@ }JUpC&%.- ,/J UaA2*Knx>&i7sۧA\=E@uӑў3Hg$ R!f9g y*U>9 0dI@';ʹ L )2TR1@a(g)5 s@]?)[A# N&Š71>+*Q  SuLxgf?`f=}^ܪ_XR(i~8~ 1Yܸk{\^T7!JY*xMH. "k*įUU+yPWQl@Xnˬy =au=5>RсoQQ"RQbOY`69v3 he*o3\PfJ pSĒ fTPR(Yf58yWQ&Uѽ0IRꞤ4P.@2^ug%hR* b]&I&l0X mP|>͐>.;έd c+ ߠ-ZNT$CX)cJR4JNcTc%S_E갃 ڗtLY<+H 5Ř']kNj{V]y/\>/.i/؋1dd3IjAU],)WyeppFY"=Wc؞ q;9Ѵ6ȽGI[cs "@B ҷ  2%ː^-we 32 *`5e<u-Ȃ( ,-~5X<% {'xt$~vsVr<\)GLMkGVKlI$dͻrQ43iz$tBY[JV)[1 J)NKX]5%"C΀PsR݇ #(oj?^tV'Q(1˚.>Li|]G qu߇O{Ezpa/~x&%#v孷3w}3_~owȁuNj@O_\q{gorq)~G ֳQQ?ͭ9_~la%VUcUUZ58iUYV0揥I T;˩Q%Qj^1Rsڼ#%+bL0:⤵VSqآE.l&К, {SGɨQؠcAę^dM&ΞY瑖a&؆uɵx4[XvvS<\~9:]r}GvȥM7x6B?〵4QHTtz:(>ōP Z{VhOqԪ*/Ubrz%6E[4W֏wC'~224εA\XZ =Q r2BO8z6;\ η #qLM Rug<L4xRM >pP.MzgkoИH$@Ѿ" 0wr[)~)I}=߅tܓ[& FeHk>.>\Txk]a7k2D/5Mh2Brsې@j:<&}MqOmn҅^P txf CGmnvu=r{~u:YƱŀG[)Z([VR[W?FZeRg  Nȫ drj.7J{ߖ>;k_Ro|{ExXwDHTzv}qg_:&Ì. .4YM$RkHٳQkي?ǜu+p圎.*:;bkj9)Km+%eV uU SSi U+RāI45i-4ޭ\ V2X *GXRrfT L3jE =M$}SjTITjXWa5&i\$C|sp |*oTV'5ANLh0vXMPD"}SKR YYM#}/0)JS HPS:<ڮsSFpsp3P+pggqX_~Sj]%pUj'>Xsx<?(S>,g$qucn' 8|yOCmV\PCBGy{%մ -?.Iz9 UjYx*nv {ʸ>Uօ!a[]Q{fOm<9Uo6de5)!b,g8)$F9孀o q'kc"*_7FJ1f}d/|8!|z VhȨc!K2@4*E b2@SmT 2*QNs9!Izb^@OW >{Ц7) x&[IH;sbZtGƾg)7+އb~uwfeb?m+@ļ2A.7~F.R;)cx 07f`JDkɾ%nѫY y~Օ@ XpxEe8Ri]N^֌]+Mf=K#[E,*5~t선~1>ExoBWGf4AOʾqB}+I!zX&0 9E =(7`lGEGM e<y&'J)$+$c>>d!Ȟs U: Yq1'Nf][¬Y%3ej*&̵rS֑bSH|'bJd85tԜPh]ŅrT&ֆ*6Rblg{E vTM` F,0kD҇B)*BzGXÈ5Tu@pr*rq &RhCHTM}5F^˛-_ MQP4 f&j-nxt^v13i"DI?ۉA jW ia 9&zOIF3Y1~T"r0bF<(e/Th#dCQ1ӫ Հp4*!#*&Ea2H] #'3m1i' pÔ])PrDi4)I@2J8!ބkA7s.4Q_8Lg:0'p/U"et=[p!bDݟHJd$&WE/NJ^FR0q~\O^`H.Id#!c^$Ѯ@‰A.i RvL7JZIz[JR, A> dASAvT1۷5/Zo Ǩ{׬"Ds;W([Vhd )фɤoɛHb2mיFʆ¤dvqpCޜb(% Hx?̬5Nlٿxv4DH8#J'YƋݨv|NӶԆ;~!Qɬ3uYZ1 I]lwN#%JPM;2' ZEŒ)zuEE+ؑR*NjKE\H>2އu):@  kkn#7Mvh/CUp8j6cJI^'~CRC @P}psru0:pGD30ֹ(8f ާH&w.ס!I@wߧk"TZ;:%foKcp%jYnw^{sX8wѷj/GcL"TȈäLGnOc^;;WTi91^,g|,79Av&:a=e? 㡺F}`kNZ`Na;]N41=a.;rI#n^#e)7(,ԍ1]gir?3'_( QfҖy)߽8D# =&C8bꄒj&9./"q{LyQ}Ì:'uӍ=i'L8t$ .tG9Z^rg~Fm2O gVD*%eE eqc ߤqbT2y`\DTYbl,4D@eIKk`*N3ZhK"M0%Xd SV>.as{ h0Xqb{E4\G4c%%H" S>ISlQ1H|"-7RX0-є{)'&z!C)oܰ =3Փa=7 kI][/?%[S*yq |ToQ?jkۡ76W`|`l^Ȃ(a{,K %3ʕʊ4FgZedJUu^CzoNjjG-d:5ц+ fTV6䮢zkd},?̅rLmVX-!: js#j (mTr5fEn>j>->M_`+r_F5};yh4ϟJ{[UuJ1fU57h] 6}.F"}F~xq;z)%:r/ԍS_x;+N>n}Q&9VXn.DBiAJ $cL2sJMDTr)r0% RLR 6PϠOa*<7Rw9~9KL;n)fA$FSfg,3KpBq(˓"Q2*U=Sl\ {DQBÈF{(lZ3CSghLϳ$xۙI hey)\)$/JVi% ES/ʫD( fЦuw8%Sh@x۷|Iƕ gp}Hx~A>< ED V/ gBi8o~Q^(\n3O B4 ^*Rٲ< |YSBEE~ZQ{H@8RVš, BE@az|b0K:\c@nIQxYR$ﳩzV{xbTi,@g|x4E#"`G@~0O7Akc\g-ʼnV@E{j88DfQXU|Q?%aLeVIU^VBBf%VJeYJ ',˨"i$Rayzt5꾦VCQׁ^Liň<)݃ YcƠ_' KLNL|4/`ή0LZ η#iZU6F9aٚw >&$}km{RSH@}G}W֢AaL 7}߈Q1i6R<[@(KTyBa2FKū/tZ(W6ox{gزcYG"ձU C - 9$EبY\a&}+ia,08*S\U =WTs-u6 Wnq;GO;ׅ۸b/շ)F>HuTR[v#.mU;3eVU#~`l7UY7 R;SwF:T ]2Q ڂ{5n3.RJk8NJ)R+Ƒ*)H$sQPq@'@h)JBPqZ(vKBRDpkWzX[罶`R*BJQ}K5g 2b'A%FO}1Aj>b~)vl\m9-& [Eub"!>Ie'T C ]yvژBc"^+ F4PɃ1 捯qeE}OY|cP!s1ǘ?WscH炨`'3L'7Fdœz%=:kgR'>X4C=4bo_펅ׁV D?LΡFwXRDbx}u!H0; Gb{Qq!N"VK:htp8 )oO{ _᧫˄^p(8 D#ƽQpR` L`( HuOga镇{8RrtR5%mhztԛ$ 3Ox7 (d]oԑJz-yv]&Ǥz5@TZ)7M񧟞n zwk"jh-nQgm*Sn?擭Չ l~h4wfo CMyijJkdw%b1pkX7yw*Sq8$eLo6Drւ5[Ndq DBBϝܴyP!@챗4]uwH<ލf4ˌ.׉TbڈO9b1eVN$|^S}s^13C|bjNQue#}*PIr\y}9J A(H KsdH("%<6vOսݻJ5+r+T mo|oMzgA^\o*1 @?/]^+X#>VHwy8'y`xF O Iǥ7!`J=% w6H`0v32FҞatJG^1HˎV۲#c*4vLQ۲H7"mզ&iST  XRq^R3A&"U(Q&KY)Q=TT!id9eCT_=y3tO<R*L eL*.#D2MB0$,TuQR2 d5 C{Gc`Ro[_dPBAygG7%fIn w\PΈҶo=O+ k-5ȶoAT1;!'DbԷ~-&5{^J(kL^h[ava9GNJ1" )=D.L5)=G)%NJ 7x3ƣCT_RtyK)aVRP8W֍+AJQJ9R ޤqxJ)NJWTk1җ#5`Y}90ޕZSj:ؖ· VQA]X =S~G$ci<$ )%6TݝZ*N|&>}MLׂY،8++=z B"Bf`ugurev6\ö WdR Qp:Y.9Q,*8>8)jtqK=ot{fնOW81*:xOyJ#{(J;z>yzO.D{osuddm+~99}}=n?-{w!e,ɖS-i(pf8 ^9]yi[c]]@Ì iTj<>݇ۗ2>idvcNrЙ'>⊚=V';5j]"P:Y*)7ãgGZIivHqVծZfGQ^\{0bR Lk+bo,gGS`[ o=ٙ 6J\v {.D,[J{=JBдv$gFr}m>}\YB?0,Fk;?ewv֋F/#` 6;E^Y+Vhc ̞%=߸Lg㞁Mx&25.Zz$N(U#+M\:?kG Pp>fSEG^ea@ig$\"X- S<ؗ!z#~-|-HNcN!8- ,?zV6EuY:svz u@}\G$e:,7MsJ0%y՟ѫ"ew?-~Ko2M&f8,xR_o_Y9 ۦBI V'd@.Bxn@VJu"*ub0\Jbb Hq 8(F!Q)S巶v[ fJI6-$i00yo5g`ydid.yt54V",67I^D_Ga5ogO-(NӱPz15j$uKkEq>yYiZC/ُ ޽ z_etҦP/ ̌V{ڮRkHxnEY? r n臲Uz}߇pdo>l޸W/~A,/?IkM n!{]^JGNL atLVx f+nn7n; E$ԉ f'#͔XdQx qi~ȏۓtiݞoѵIku&BWћ4hڂq3p1CdoQܷIܷIܷIܷmqow U _AC 8UҁD 3QMղ61_Jj8U[*]?p}kJqZ5c^ԭAG+-Leڨs}1}Vcr*YfcM*Y֞ōmeY{2i&d/B)Q5)rkZ{gPUچw]=L&"(=Rld gp|,g"Ӂl<Ӹ.z8wM $jS18IK\4id/Aq\1Ud?X\kv?eo4>+7Z*< -:l[oI~?=*`NXvmD]T*~7Io6^Uw-3-*oEUF 8(U]H)OGc?;w:: כwKgN^7M>C)6Er6OK]ą8-. X5Z%jM`E-d*fpkOeBGpBP ,4fE-nsdOnt҇R:kSҸS ]6ӿ!=Ihpp(g+aLX^'j:99M?/>@[UAIsSvx?xtfӲZ۳;g[9ĝz}H>MIF/|:i_tM*t}: 廳r9?>ΡjgfA%I;xIniw?y7䩞752N3G"`\<$;~C%RFŨ5ggr%sv~zRΙr  lݱ>ex#4V{{Ox{D22MoObx8hm~Q+3?X1XV9w״{ 3䨴'_|hGƠQ{sv{@ -/O Gwh Qt-Yc WL kv[}Řyos/ ~>ܶx['%ЬroAʞ3re!%c|C"5Oot>d~V%k:>`B"# ivpFlٛ"HbϮ({ŨxF~N(>+݇qv/ G 5}lZd`PQY45ff-7τ]fZ `JLUh9.N@rtw( Hf,m|uRpYS|U}ZB F'Xj KrT_"R -*ISP )R)]rR`G0y_;Oʐ0vX -af;ϵm.EF=i!`)Av%uZ>iqQV;ύTIj{gVK-;lTe}/VYDdjiY[xNDQ}K5(xyB>N%-KqZuovVZѬyZ 1 -҆j-YK/QKi)$M6T3϶µEXo7үQ}K5m%k)dvޟ?~X!Jou7EI?o~&m ~y1+KlA(UBBJip’jWWqI^T~LW#WX59]]Ur?<\P=zq)X7EeӪjn ڜUr!uT\uY>rKSJeq0H>LS5ZzG`mq3"{]ӛzDL+lPs`̟J䌰dugp|,֙:ӛ:<Ӹ3z8tM u$j*Pu'],]L)fCznKq\:1δzi oSQ(>])< =l̡L2 V Pf{֙J)*Ry[fE*NF+Mzi֙Rf_znaH :ϩ*͂W^!Pw>J|Yyk{Z:UC5eǐXX Bo/2=]6svzqW(3BrCNFlu9.{rg@}4Oa jVkɔWJ/M߫7.EYQdk*^y͖JUxg挳7pzu1'l~xND_W)O'{_X,P,>S_UtΓPJlVl`ֳ ˕ʱ`f{u13O칌Z6jX` 0jQfx2V\jNe`AA+XVeEm3E6*QGQمRB&pKK?jܾMVc_.jhӢQ:tmD]^_ʆqI$;>\Yy}) "z2h!0С/ghػ (TtSv̚R ^[H9F۷ݭR׿7ޅXb6N9Db=RsI^"[`$.|Q  e(W.7eX:\"")xک,;Cp8_,;e/A\Rɥa]&UmtRߢ K2xlʆJSUt47i;M9ۥN2AaAz&M8 ]z 9" Çe"a0ܙw&Ml.gwE*1 RJ2V3L#wR S*#<<$^Rx` 'fq4FedzVPqv2) ec孛.Y(yD_@}{3Ғ]ڪa=X꙯+\ WIp -ϕZqx&ҭ ޫidtT ?I}NM }gph#f}4bybp6ChU)q|7nl{|``(ͪ;6yxrl.+Dw<v^TvУޘÅ>mlf,^V*9/ `ץ `Xm(ca~)) *ʘgL~RkeM-I,?wv$͗'Z`Jc=aLBM!h =y*U4 蕠z4KT^&a'$%#&J'\duH^ Q}aUI<9h<_>)z}(Cz5@swW$tŶ.4MLMB5TLVfz` x|]n!O/aq v}wA*^ţBqO^B4P^W%? |!eQ׵/fohQITh]Z1RD˙sqzr#hBך 3$at0%yĻN r!t}]j>,^vˁH_ ,PJr 9 tRs#Qp"bx DϺ"@dԶz,gdGjT}}n~Ur_\_~fe9H!bbhLNXg=H"ԩ8wXE{PßaCT}U߇wR D)KnЁDXόVJ+.cgEb+aeu` {!m06g 1+`L [n$2B5TTp^[ГbMبI^[C"BO mk[obE[rnbUXraEdݯLͶ ߪ_~xx5k` #Hߓ=|~+"rwO6]hn"@4z?"oU1_,׀x?ضM8F\H$LjKj45&ڪF'c@\jּ'}B&b}1X(!WM䗮Z7D^h Ԫuc~YC[/-9!"8&Iu[Yɴ!8'<ƌJ\7K5})bv5MC:Ԋn-qY/->C/Ejh~mF'Ʊ}G0d{H_DS00I[xyKԓsqXc<n}e\T4vA scWA-5 mg%ADykѳDP`g"Z q_`-F,дZڇĻc,CV1M L(Kd) \Y)Jk6XɘiY<0!|,nSۅC5U!v7$Vg`CckARx^Sk-Ă %#R)YmL``2%*4\ҒHlF0ueA[SL R__ܧ2*ExJyG*6_Ѹ'ˉ?z=o_ 7sf-95Z [t1#O~lσ6O[iIjh3Nh--$t3NzzpUx T36T]ݒzbUs|}|2 Fju{{owvu=aGףΖ8o;USx/WowOnՖu_䎱i6ً8)x\5H4ھ7ޞ=ՉgX'<|$䕋hL T)AA EtrhNf(=Tօr͒)v-);Fv£?i!}k쉦j.$䕋h{Ҍkrf'C;kOv|aϊ4 v׃A7lk݌ \e3a^"qM j>Fhy@ Lԁj1X'VRLgEԛhߍfOQEU04g3߬~Y$oc LSDԁ=esϘX*jl35\6W3P3NNJKF=i؃j,HKaPg3:*}u.4;p((sc˜f'MF3Ti<\LΗ%70RD80CDC;j iD44d&SLN#kʺIbc3hT4`LN3LI۩֧-I-Bi'䬞dtQT89kS49 iAѠJԺ<nr6Fs}cKo-%x漓ilsZcο"VݖiCE^3%H.x$-! 0<#ibJ8UD;dp9' LT+1cvi$5TJ["1vKN`"𔳒xIc/K/@Ln.ADҋ[DS[EtGHL1Pq(hs)o 6yƜK]5~"u On~@ٛc?=[߾(On P{OF%[O>e@UC`\Ks?/W +ꗫO] #rVIެbSlnD%.sQW[BQ &+V=ްXMa>{p9hjQ ܤ`|\'_R˂pJ"<5?trlcK"/\1YQI/ߓD)>7LnlomOnܼ1YaƝ?/S-~uۇ"h_SxPk0zt8t?LpzUfe_Z疸Y+;ՍAb1x{SMl{`U95,Y2JOJ)꼖 ŸVp~gw'*O~Vxb7W[D/Lf  aiJ61ګl،X??"S8`ի|e)$tRZ%Pf$x0$8(}L`XDJf)A!{wdX"٘JM:ٕč؎ ;0o` Ϥcf̉ł 6RQ#qi G֗PTS}S]9LcPA2ک.9u7ڳ|,Z환\K`:8M5sN%"ZHDTĞR],M1jz ;NK#&jxi60zě6~?jciezr*HwG@עCE͖DU)ŘwSJɒB(LPhADa"ǔaoLxukZ'[ j<($ӰFHj]KZ!ZąPM![i.M2D[``3,ā<#%\hnPƔ4c|o/F:<*CdʞKCn'L$` 64$DrFH+pol[wQ[ئ $w=_M3W[ FOhEm:9lwumDKu-[nBZ[=6ZJ>=6cX]Vi!zۍS8yNA",r{U o@s0e2\0*9؊% j)ш%z,9xrASH-XD#Nyk. %iyACFL¿xz@ 8(^gg'U]9JR;mxFf*\0P J uw Hhw z`hIAN۴ ˿e Qy_8BĦ&Ho[c<ܾgav {#2J#5`Z),͂6%S` +$:1!Vc)P񲎛!AwmJ~9 Z&wcsd'$'/Yl6{,DId }ɖdn[nǚb*ȺȆ8qYѧZ ؽ\QD{J9 dhcKC MhPnॻĮ8ځQ )!Y-pb%#,-C4m h*R2 D"j"eƀLḱsk$2*A@1E?*!  #tBRȑC" JjXL%6\Y~ksBO&CJPRKX* "`2)TQ{V0EjN00S>f 0lYEO.].C$@Ǿ an)'SbDf"E̥Ǎki.4ju(Ol'CZp[p)b &a@E`>?A|+TәFp#޺X2"ۀsg^gE`(7|rZQF>>~I׽/@12_} VwiE0^pOBv?~a5-iN${~ޯ\O>}V''.lj{OIN7ӻ+g3Lg_o3d|7 'P.=^1E8SlGP"8&B%ǧȍ'o<;/w}N*:#? ,ۃRڧhO H`iu[fb{i8"D 7>7d3hل!sIZr&!NJ@Ko-u ]Le|5J8~b:hZuyP&յ- Y ,nɨ@%J 䖞ZO͹J݀ V{QVQ@ʇL[վ\|u'(Z[7X* DS\Z-_;/*C UG=J^Ww7U.z/+XnP%-2S xW:t #ǴsZXС0S4Kit:R&յwC5+m,!BcQtCx!mE4JT9-щ}GvFK*5ꐐ7.12Źx)-щ}GvN/5ꐐ7.2QLaNdͅcΚ_*2±Hsm&D.Eb/cNdlM)? ac2)܌ șɄq@؈4Te,3i ,7Fj`ZUv=։&If'_)yx`6϶X@<^y.֭#V``>H5}Wq 'O2l,!Nxsa:|nV4i׬BF*=pT {$kcPN~˘c^b>YxEi)4`䷘K]˸gPݤM=A أvGǞf|Z|3N.$ڑx64Jubk:=mk܂kHös>G= Lk,FWz~ev|zP!B^^s<ԺGlkpzcpT9Zi^:3Ĵ^ ^w %].9wi)!ƛ(/3v0tGͼg+_/(i}"_SQU0 {RLv/Ki/2w반 ń.gSkJ /Oι擣Mױ U]Z:w9"_WղĨl6/}DI TJgm6/RDIB|n\y6M.'G2-95b(ՠj3A0bum mQ%Z mzm ;ȯ)#s,RōYRr2✥ZI$Lc`8әA0H%Ky@ʉ\`:ui=(IJqa8P86j Dw-Bl2K:$LHZ`=gK飩FYǰU}\7-Yfw?:FE3Ma%$M5bb>'WG+sĒ ?׺* 3/ViqeLb[ [VE[c">'Sꘂ*;❭Z atWL)&ݡG1Zl:(5/ɖ3RX ˹ w;'ߐjgWN|',zW' p(zFZ0 0cnd3obOdG8x\ޢ8ivlTDt[g3|3l:Tz^SK|$5(ۛO;_O?W>_ՙ+ iuY{>wseb_>E%R:sφcLR;u~{/^ߗx9 i14F8YfLneBF gRb3(X&{|,XrV{E#d^nz˦ѭ^l7 Y˷CJpחzto歆=yGIU-R[Nw I,XVF`0) LF。 ISZ Tk5VNc@.ɽ/>緥vC>CZ,=.eO sJ J0ܽ_}~l0v4xtR?N3xHR?zwt6_燧)}q;}F|ooq3P.[C0ʐ3\ꍗ4'b .9k=WmC+x86AAGs}L2>aϓ(k4'f6uۉ?;Oc(<Y>ʹq:8Oksl3yYcYILEqт D >}Q P˷ 4;尢@o+ Y!(nO8|H ę%\<~30#,_ ǎY` ,ٸ&l`H0쨓 Zj=Mz僽lP*2`HL&6U8=g=嫯=!#\/߽%^8zzg.7; 6{JߕfKmUTa\c7W$H ҎOl2HO ^[Π?qd%qZk 'J/ )9ei3<\ yޓf1bK̚Ҹ_Si tYDʲe^s>EUMq㢼9XJ]4Te.5sᢙ84[vHU8"U;Qș%L)'/D*UaTJU8rLʚKSH P0i eN<P:8ykݱ[h0NԐUJPȡaL23 n 52<8-2r)HR,,A13SAWbNr FFj4 GM2U2frkfZ8q(˹smBsih ~V@φ+ǎ浘1Q:m&bNg)aZ}qs41( 5Y}/S)LZB8 , Ɗ}TrZv0IRa"p(@Xq%?TL3L,$ `x;?ŠEĻx|l r*2*c]d8[6 Dx!9%Cʍ$U<ߦ:Ѝ 9cj}:sϜΊ۫:\JSh_.5O/-O9[ZuZDMSF=pR)~4\֍Rlz@ӧ)u"nΟP۩톶^38نXEST4^rקd|s}mQ,C1[8t)J<)>7 3I8V?g$ǪG=Ge4q*pS:O,86eGÇ8֯~)uyS7pyK9]i',nɨ) Gwlm y\\&x8@Zق%MUQ֪(8FӺ- DI%vv[B_Jݷ!$n{0e;OEd)`T:DuT乏T+K9&Ѧe\LDNTί-%S PI__**P.{*su1^M'fQ+jxm̞ٙq?0[Mkh ~پm_}:;fTgWjOr.P!!o\Dd\5vS^gnN;h3z #hOg:j:$䍋L@}dOvdS/pR3&ZO~jbDhnvVQv̓Y~Sʵ7Ҏډ?)l2oPze3aon&' Y3Q9̫B~qmȼ VMWFW+%֬Eٻm,W\~M2Wa㭚d&ս3y\ 5,= uh]@"U3iȃpsA`kfA@ D(y(F,:z篷Ka?eȵe j\Ќ BwJ K [S\p& _WZ_|ɓB*&k6|BTIbVNe$l wc$؏um 5#vU^ *2"bf]E`%eeXwo wE6R@7NLӆy4M-΄l;P'ZbgFα~ ;ґϿfɓIrwʛWhIrX΃7Wx>fY店| wc<_عd,~Xl췻;~8Yӧg|%ؓ#$I~}*5}yp䡨Ox(F(j Tj1FbfD@QNcrP|kͧ;Ur$88kyr*{@A8f?P(b5 &L!Ņb꫕qDr5T6>)E$qPq48Tl 1 #:1@R!--9`K$.-TVf$ Y,pWXH- ]e?(fgA#9ZqCL0(3b+JsoKw=K!xC] Ö5i5ж j]'@ z4vaSU{YJ֎hm#Tuq1i%uVN {hS(EŰZn[k`[ Ve* E**}*}ݝwbԯWQ1kJɳ^Ꜿ-Qׯ:5}d7^Kk0J_\%`%y PLھ}5r)l-ոBfs8FLiBݹaFqsll1&b e;8` FXHbEP)E!PL_Pd%8y·kYD6B6AC 4 F!*\jHdRb,:.8^{^Kgcv7 XX@z0~vY$܅sdvQM"}5}ƞa}\hWJQUU4 }HgosV8A~; 9'߮gtV^;sr 7j ~8yx{?wwn1~¨d> cQW~9G%zJvXs 1  exP dB@E zG6l(x^o}eusGm^5OH&w%ԴJZ ޳Vm:?h/ƫVjV GX?[EG/q ldh|}>BPh4Dh 1"c $DŽ!`$T0 Vkʅx4^QAZݺ:c~v\Ep:$}Yj%.ڱzPaR^CRߗ&H6гv̫s@,`׎bv1s0xd;밌eG}/x]m+]X_[2~{]iÈ(VH]kfv;G~'cg4?>;AztR$-PX!Giu݆V'Mj_t=d] VYVZ;@RT5s%@s 8B '_.4SiѥKIL nosː(fbT@8rG[]r=;cDmo'm^yQDn;mrl=ASev.Ni~`$YMȨDڭ>L'`QuaNg0Qk>L Hv aasvl0zA!;DC"DL PhcqXRY= E\Dj`G0&s"1Cy RP6Hb*AÞaa$ uyQ*RɢVE R񄎌,pN2ökv[6T_e Y3, [KӥRfsmi$ŕV{#K,cKVDRŖ_^?"[]wI-F#W1j}$ύdDF~Q^+Ps6/-yfyV%R2 H3,ZXL-ۡ y ,:+C=#k@.wJ+n]ƑBIȞA}Y`HևT/KM$Bv/;q^v9KI}_1O^=u)#w&T0.tni iOo23\&+E1R[X}k͐37#hyex!3Xs"v{J3|lJ1A7p)ݮd6l 6b^^G&V/EʹU:~ R=ZKT;ϵ!rA\ƀd"j #);B_lgRPCt,8e94edxКU%\=7\!=0Ji2j>X ho ܇?vWy;q*q"$ a@Ap*!("^z@V)fLߓ-|qr[`o08KZ&ܯd#@?T\C9URxRIMd69^6ϼDe֩{rh HF?"d;.ᔲI ѹ"jR]}rØ87 4B dvR GCNE0s!nvwmnm ,*`?\'"ww}h0@ h?XhyA9gӀd\Ģ?ʛEg\sGvmGvx<^|,{u^H^ -VYH.'iP;_~u*fȖga]tLN6ؿ[0^[`!j9`ǟ1qܶpbE3q/ص[q2%(ErĉI?g ;Gwy7VFy1VOeJbzF\ĄBJq;ק $h%ADP #'"5I")! 1ƂjPµVl^D,% tAG=Ch I?+ :r8:#$-x ܜƍD^CR7MfYmÏ/ɟS8s9#/˅~2Y2G[;>kIlힷe ;̗%:׿͢qX]nE/7<ߟ-gn8KZNw?n_U߿~h5Fze$t|8]d?$}D74P*yvIjn\[з߇u%̾Zޕ1Sw/sG;%լnsAv2WvJ>? b:@&wRnS2wN3[U8ݟ̬&m$ɿbvv$wWWsgqM9M;Ѯm$yf&%%)R8bSzu=|\@j4/]a⡳o>M\6G_}(kZuwo_I2~EGW׿DPf®LH< >xEoKJt|=Yp)!U1mVLrDAK<:{Wb~ƳK-L^1{ Y~od@oeޗFʗ+[~Mg\͋Kߔq#œuF}j+(ZJ6w\YG"uz+= vP="-H}kV_"E}7ljfmʞ\WpUޓ+p% Og9g']K$OyyTG ǀ9.gktf&g$a]&LMDM&f/DR\P0.CPY2(f:J!Ӂ9X"qPάf/܏u<*츶^uBG((/Y&>0R%*qrr;u4|lu"j`7SXDڌ1],\ye#YUdc[mSXxhmE[t+#D;Ƃ̇KBYqXӳQ.<$ݑ<ɯ{1K,M4.f%ݜ:ɯI~(*N:fi28rVq:w f:B,Wst'8p6$_F-2_x2>AjpQ'^d}w7/!` K|h巩(U^]xwELq=9O'#,n<)FA&e$A3fLc)mQ{/_ܜ-^iN^gD띗S$ǣpNsK@}vM"!cP) e!L㎓Z3*0'sУ.0/T&!!63im |bP[h,ch'!pC$W'D0C Q F< tnQOD2&Z@yV}G}Pg57&`y#qJ%\T$|J S+K|jTχKj\m]kb'^@+7LRՄjq\ 2KK>s#RL*b/&>sR(KeJ8dL9@5LKRd  ܰ,pACx( \ l*cyL{P)8B<&'wΡg(ps42_RH!b2[i{HQȤ8 .VCnB>NZq43 R{^DnuL钇n!AdpTfO `(ґИO.\Oc㪩J^m{0R_F`7`4ķR_zy-uK ukb?GNi :u[Ԏݐ0[/'NE@@jbID$@{;O_Nyƣ}ieK@ѓ ՠS㉐K ?,wj} dQZ)hR9=,F=Y^@1qnܳډ npoM^O~nhD#QpN:$bq]䤧+ێnH__H>V۽mO[kAsi) gOX)qv̹ʕm ƭʅ<3eY3 yCA)Cu'Gkk F?@rcaA|j*'=F`+0KHO\#)'jSIeJ(}҆kvlʯz&'5ZDձþ꽠k]oqhaϺM61:D2cu+5S :)M Y Lȳ1i,9G u0*,JB ƮUR)<϶6E =pjUklz_8k(ZJ 0 Bƭp53qV8!sPyFkiFA#&1EuiܫZsk4'h2F;ςvt,sȅd4>3T̆{vc#H VwcklLġ5VXt;' bg(Rs4AJDtcjE AȀINXm6>@A Y@4D4rs6/zۅZ#P땍鈍vT[nHX`OR\DڡbjTN&=*hek,Bʢ Bb-~lY;i ա \ab^dR䣳XzF֘3$o$&C/L0 ehFnT{@ׄ?7 C$Et,IѷNnW>,V8#)y`tug ^o;Ʃo8@>4QA){YWCР"Y؍ ;=mV }[%a=$c#Z5lOӎmӎM=o~o\"Mu#~'kP%tOզis ڠ*n#j%?\SjE$wS]yuk#;E򤛷5/}} oVPfԲ ?\ Y>}uNZ7F!:rSqH$GT55!9=rny!Z# 6ސ`J>:- zO %$f8Dink&} Q՚85}Q'u2\*kF:r *C9gsL2rgƌ/W.j+a!eZjRC}Ѿ8gYC=<2FO˓? ?:__ՔUm"Ymqˋx}r3;n盝Ňc?+?"u2G; n+T5BvϯqvUI5}Yj4H#a>y xGh+Byt1H8Hah"zJNʾ[D|,ڊjͮ#݌U!( uھtӋ0B[ҭ YtgLa VbRvmJy65{;_Ljrk=k,xhV6}*DBzXA";%e5 aA:z|rtf0Jk@Q)96!2]pTW\Ymr5'Ȍat] !,VO!7㖎p(8xmJC y}#`wh7-xgD4#C[+0n;l2aj OL`BY7^0qBP;!Beڄ 诒;U@yJ4'GmF:6ܐJ3ys+ ɕ(8l=C,tYX N|"pTϡ/a}RUv,mE/`~~~_nC۝ztUrZ0KGYyA^hШ[ 3ƥF9#w+s:.B4T.Ÿ {A(7ȓPR$ 3HUQNl#Tm@7tux/.X@^Fm -vGj2Kc\c.]0'^Yt7.:C{?8O]ucJn=y꿯ypt᳻?}|)_ ^#v̀Jܻksu~~=8KWlܦ e->Ew ]f 'E~/dϩ?ZF,܂cyg;j'{1UׅdbŒ:Rl?+S~wto?moov%%mGM,<YkC[6^Ga 4sHp,eB'92gy+,YtYJ,)K)umZuu؏UD&{4j=d|iw>,+ 'TB,rʑDHY52Jr.+_+F#i 4ʎOvJ/@#~ H{tɣB86]FHg}Sc9Y~*49oZmxebP. 1HTBcѴǸJpqҬaopmxv?. @<w1 AKU bKզߋs2@0jW`+v:ZyQ59?}r rJޟV¯sWQaabfe"DzR}ۧ#V¢' TKߨ{ާQ9[wK vbgQ78N"-*T(u&wUHr$H-r~hAjK+;o{ XFx>?/oz$$~A_7?}8}VCG b Įhnooo# I?7+l+Y(5 kEh\Ƌ43\|n w Қ77Q狎*)W/DB̾JB3X5ۇF96shD (09EdRn49eT7AbgM:vBy#3qKxԗ1Žm#`ӪNcD}*P-yihOpT8e:$'< 8&/@ y}2>L oWCu ħ֔?0]נ—<VG1.7yqSݟpnt(jS-F1d)-lMZY.~_D[gZ^`kyM;.{TU XlPRZ#'w##ofA!Cxn?ܝݘ#[:!!(MI.Q vMt: $8z#u^2ʝ9rGo~nח>'Y@5wp n#qh5 Rg{zjÇj넔FEEq͢., 'âM>=,E2'69%@ўXAd> T|P6'MUÄgySQ^KfHӡ*95uGa~Z 9?W$jM|Y 4Bd9 9eZtr 9e&(ςQ+rljx; LhIj>6DcdT@:&$hHRj~훘,^5'NS*\"hϖbMug$|MMh)E-(3QK"PJGɖ7ZmچAoVԠH zM9ރ~c+A+%g:81;t v 7 (iiJyGԮw1#1[-9W(!foqH:%11[}I6|&ZcS/|OixZ rL%m`&ɼ[}YwBr)K'q>nuѻbc:(ns"PtޭV ӻa!_)Ǣ.:?w) {.Qf)m(щ h)hVԅctTQy5Xo o^ͮV+Fh5([jR&ShE"rh7JaJ,,˕UKsΓqzeՍ'*haA>+I; LeB[h6RaS)T,Cb g=Owߩ.چZ8{5؊P8>2[KEhs>7iRvo ґcfn4'@ `)x9';b/0?2Ʉ96ռEt0]бlM]R*i;hS I7X|„Q- 16Z IuF޳ jSJz9ctCpV-#r402FdYӈ{_cҪ hi@+PVӈ>te4Jvɺ%P^ӈ*q}֢ֆh+xi'Iuiٙ^gM~~dܾw!B'N/ݍL9qGr( iJ맏Bru87EyqS$pvJ1J]lFRhտTrRPWNJGV7 E GeC]dvtۣ'#Wԋ}'';4 @xH`n^#=qPP\E㺸:J0ك}yO<(y&y&r$Z[*;\P(o DB <鎝jWLUo.ZZ st:Z*=Qb_2/)badqPKӛm4-W!YavN;Vr@m" Nqz]q0"p C[^; A""pph`8Wk-88hd]bn]o۱Mm n7WonR$FgmieD֣Ʊz 1Xɀ嬽+8:k/ :BczTr5&cM. cBlႌJF!2@6<f[jIGo>.|ZoWH7-b~suG۫r^nFëU|?&sw߾YbP,^ Nl>w}zg/~x57eZ)fՓ+_7Rk bglW@zn)mJ?+{yҳR򬔷aIfWjm⟷L_+ƞL_W+q\<+5zi~+JX#&*u޾u:ZfnO9BoafC[4VTV!5zmW0P[soad14JK'k/%.Ja3_enmQ@XB+^ .d'%oXYh ,vAɋ 3QeEGtf͑+s/#p,|I[U }cY8uV+]o7W}9.zGU4O8#>!`ݶ֒H$"3#gF#ݜy&["Ȫ__5똨֤ x68ښspT (G(rUfN46&5st!; ?RZM bf¦uІWʇ?2\'}^Y0XvBY!`v鬐e!}4_zQ#N/dH/mk\AN-(Ҩ e8oŒN=N%OH`1)&l4HUu  7H.)` Z>(A  m-P++PxZzlUݘV<LyÓx@x+}gL'$lq= 5']8,ZY2zMR9cyCu?ձM7Al,m, ASqo,2_/1KrsJ+MCZ1VR\ #HĽtI~j_̑ >0V]]⮚jUNŢ7Q ں{8mb~&8ى f5,dTHZ?8$ Nv,&0^v6Znl!F Z}CEq &K+[9/e_o#&TH-jr+S~RO1H%+c*P 'TL]$Wrh%OdѺ\A&֋Rl<aЀִ(9.ג%h y bX ִk-pN/S يZT3 jk5%n,jXrMCF4FI=DH-{'וAUQ3E^뇇[ `z qߟw[fW9T1Yu?nRӏWW޿;˫?g]+{7 ђ;7+Iӧ|$)"G2a|wF Haٯv>9>RlEil 3ރ\]cܹKH,̸@)gus(=7۴P;?2"bq4C9)o{݉ZjJJ+LVû0YKXLͅ:2/m}=ۖ) ݽ;^ \>=4E%N5?5?ږOC@n,1 W{D rlb3[y &H.I}A&(84/ lu$iP0p>lNyÓfLz7]TfЛ(4@M'{ú2YVZF*fEBamhi^Js}6em6։T1wk=\[B:Ok$r 6Xu!V~dM}Ր&JC3Kf02D~XyPD吣9m50UޠWTӁj:9T+hYRԎRU;X(KWw,ݳyEuAq ^/T\I,MEhq961zTX9[o *:>H֍vRRu"0eoʓփV 䞡z'ߐP8Cz6đGNP@JRjFc\+9G$@:b{ϥM#zm9nHIQi*[ZHKU 7cC9>q.s%Pg]3@H|Wdk(f JYy-PF6Dp+υmZ" '  k剠ݭlNfs2I{)5X !Jv2k-`AyŔXKE- V:U#s[A˲m(g{R*Hˢ /}PO@gYس)·jxۼ~Ɂbw{Yݼ}fps >߉ '6@1^b3dobCŭv~)h_e-jdN R~2Z""bǮlL;]_PMY!]aGSbb%a9u`*WSb:'hȌֹ| XH#!2@ĀMcQfŧiwM_Xu_N 4Wݗ}맟͗sH/+}Z|B-34܀6/6]"1BxʣWy3S`|(a^fxd $.[fH̑3+v_/刀Kb R2G +HLv.,l9 DwJ:jQ\L!JQBL,<y8+rJ@ONJ`:M7zHQ~&hWƩp.J>Hty?_EgutI-M0 ;sGu/.A?q"DC"_x,lo1+rj$\~h迾Gab,`u]5ZL/ISU':ނd5B方QC3Ԉ-D-_RtNL !!1IkuchQpEt79Dև~Ō^|%_>/yB:{Y/˧GdMAFaف@xZ %O@ő{6uG9ІltFxɏ6*rCF &$ 2$K:{sr0ι24 쎖-C T(G`Y5/ݑ~K8Ȳ}Sp /6&֭Q"nzo~,=[5VqSW]u\_Wc,iWB2A~cIydm|iBQz'T:G]k~ Ep#~iyGdl! MtM;8TN A F6]uh[ MtM >݈"v*mSDimx[ݭ|&ۦ4qn[Uc"+8c5SLBHC\I b{!&ܗOi5lmé9%r劙=dF],ٹt㯡x~x'<Õ=J r]] }X3' кx_0g ք \7$؆I) CE}܈]9,8ZU>\eTju5ŲYl#7pI瑁ɘOJAj0lUqN[ܝGRtg= 5-:9Y a.*''9zP%V չMPõuCk1Fj(PLrIKKPrO6"Ke_(6CG7hpZp8ՕCK ϗ~hX38:BpAk)aҙ NfӂRx@T)N;I_:S󆽥3XY/Q[?~H KUKQ򄨝͢gɨ=Ţ1]Rj\R;s!u{^>HiF\u >輸>R\zy}s XƥkK:t71JQׯ_ d/7k`l>#d_] e->{3@(:p]1~uc%ؽyuq TUm7o>]^]|]Ӑe^Mam[J[(\jsPTK_-ׇv}ݑ> {ȁuiz\g@A9Yp=)$x쵺+*_:g8_$41ȅciYp}v :45{Arf:C2Gm)ԌgR8U:o,ij*˴X3GKiDzKXcv+Fq*jVҎweTu:8x#ANYR\ó Br3kktm*8r=ŶTP5`e>DVˋs }_?Xh]w]o^*?(ɃGZ K#.0 FtAyte|0/qh2 "W+pH-2ROxF=p_ʿlJˠ{>qCYPb'BtgOЙ\ gHO[7Svynqn1RLXZ.P# d`6L+IZrMa 2mi.H 2I R%7FQ8>" OL'hPFi@¸؍dJcۻH0_a k7OūVuIZVoz*̬[x럿l%"pO4%~c"~L|zĢI?IYfEwP;%xo7\ J[Rمh|}+G[i _O%Ba _}(]e`gB+{}|) rmKVw^j ZJ;l<0s}jGY> @=kSEY'.m ~mIj"T#3y8|}ט6cWX䯿fpT^0+1ΗZnTa F v:/.'2\#lvezjxމjrq)`̺@̬zPgGan1`tv=.ɭ_Qk,(i28׶*; z?Z&_LﳭbO 6NʓeKƘe] vFiTs w}WM*`==Dj.7E-*UOzڈ}rK^3ApЌ]qlq-];R ʡhR1,M;=vǦR6TR1C4М|J)x::Cf̽,M"J6 yStȜ]wgPLhnȦdmxHwc<){N.@HźS kްRV 8p4V66*yk4͎4l kԫ^Xܾ@dǽ%<=N)c"_oC㧼>qzNԺLq7cӁ6bE^+p]?f]{"^bomɱU۫M+TV]reu|esVăp貹MV=z훻q5)[R㉇hϔAg)`GR|b66m$VJ'TZ;K^ <ȘJՁ XB͘0~uQowG r/rt,$O>=fP+m=F8Qڇ$n{hMҽĚ$cAy l9+M@?S1 KYh$W|*&>юPn|X}oQx.,2 [xəIPEL^Ct|mpD` 'B~Q? VO6l#RB=!X4PHC(+Z6!fws]jP4T4Z& hh$ƻK+]3t3=r@F[V>I" Xj-n^v/iU}l^ft/+sl2 g.#nO\0l:pJX?isk5bfQ.66(.qF4)QtڠA^?xtC';| 9. &`>9 h*}Jq Zi#WZ3 [Hmp")z_uUZӠ8@]`,l5tNZ撊<8#ӒOEjNIa幥Ar9=Imq}>Jfe53Hmgm+azݹ{磲( Xj%{,BErN,GJ'rb! :՚ײ7]poLtPRISW0'#35\:5~AԱ=^VU,ܼN]l+`~fZÎcF{JSh@k!um*5Kʂbf6bpYeÂv&1faZd-<[nj s ΊC*"kJqRUS8C;4m{VnXxc && X΄uI+uK. СB%>F<\&Ƭ"3Jp/I2븫Cu:('e~З5@2ݗ>daUF@}H`B&NLkR-$5RYttJM5zG׆[a] /4wh݄M4JvkaZ*P ԼwJdĠA!Xeȸ#ݘ+m,-=Z # c 7sc}/5VP{m(K1=MјRG2 SZxZ?z{b)@Mۯn7 e7^+t_6|?v6׀~¼5+Kp7*XT:'AI^?m$qN9J[H4&il%l*E銖(5ZA|\کyG_2`\聓fUY*&SqyE>H_TY(\q.vy0hGчP] QUlN.DORF}ߛTVOvܦz>0izx>Fk) kM[fOOnb2dQ) ?v]7Ԩ@ޓE{I?P`_5.VLH98x*.k"ru[}uۦdĎ`riE-5zA=$Ԛ^ԡh/1,6wΧw̧9|~׻ʌmjoxv^Ŧc{mm̧9%gw钭݆ꠊ~7Zۜä̻ oxnc̐g>Es *X`ÊjLÊ C\io*RPQt{MN>f1N,(Ef5JDMr gI:= d!`ֹG#qn6 i'mˮ;+hF8h cE R#66{ֽ {aݛ/VC?'[_5 HT`)u ,d@J0/|AJT|`+Yi$FlF<2xߋmƲ.0L,&fR`15ӵu5sƈ²>pghdV9:%h؁ S`3lyHRf*sKeD(5 gm9Ib!*1e\ -=J=B .!uU!:u*.Rα瑩TYrNﵧ"v (.'k!}IS*RNH_7GuqyTpmZQ: ,*(A=zᤋI9bO"⌞)QrݠY$©Vr\Oʩ?*rg dJ )0wqJ",J &.[g?yÁ[P_j.YO(#^HK92ZYpv".8f,ST<!J]i5xijmjT2/BGjj\4]X@dN{5fZb3XnV6\v;k%v&Z㳖&9;Kl䚳6q#җ72@h򇜭T]%l>SEt$۵~!% G$yPfp~@73Fc8KgiDr{' w=ץSv5P@|  WV2Sp:w#[ߵF;d,^盲@z;m?cxny˜,H9L^ҧjUۻy*9y3OL韴f{փ=pv Dų`C lUm#מpUk\5Y˴hVQ7Q| ~iQID=!9Qaj s!gژăp( j)3Z2;* JuN \l R\P-U GR%PD3 J5-.8Q*"UsA*GTWJq͖s(P*ҜPz(k[R> J~:7JZJuY(Gyj)ON#G8*2-ǡKQ}^@vcFLg&D?&T}7qddZ[׆RjA5".K~ (Ű9Yo> 䒛k-`5 &$VC Ty`⭱z{>>]g|"S:S6M=11ƳknCbkZ ؞rO2j&#X)0'4V2{k fs`Jr)GsFЛ| qB+6)^iWtFz8M*=Io7S|ϋϋϋϋ2>op˛)R2zg6d$H#sN&YZ&u <2&'?XH „570j_*T `u凼y h>c2z3`,Fg伷fGDQf DF%;ptdKi8X {MVK.AH ]uY'6c[#D9S%1$+㧂lE[DzVY Sbsޛ31:}:=N5sޣy_I{-j`Y9.:92Jxnƌ(_Žɥ%yo=V;4~*!nUS}jTtzܷ(U"bo ?RM? 8nġB`>WVq(-)5Q*0QT`J u LjRbdFR㯨6ǎRR5R(:JͺN)۫P$g]L-כĞWJulf M̬!%Ʒ`DDBT.3">V6cz*S)ҮMU.5CJ?1}N.Bbxq)롙s"ʯo_q{]+願CQ4FMp[,kzyveR^<ɦlʕF-zw# ~ۋ:yY^WYUr8db6,2:SC6 sk!lJk/]Lghm"4MI:R6,G8Www\PIsYр8oFGҫC1^YRZHo3Sg6G =$37: QfLq)RSRzCHv$,q!ݰ]=Joyi[PRa60ٔLZ!sY3R0ZZ`63&5y ԡU"QxLG0ʨh߄Qۖ`O6! 3Z18 l@έtHdIΤ JhjOac)ճ5B.u]VLyN 'W?9i?./fOo":?W!DifA12sSԤU•wpZ:#r:bK1,s"}":l4Zh!5C\LE-D)=13M9KR>K@Ǔk`nτx(H .6ۋ>/yvm}׺IvUh>[߿5B?7Qߟߟ~&Zox3ϋ:d_IbbZ|\bn8ߞ\T:$ B8\%w}vP3&#;Zm%+^GzWؓ [އs2;[u'>+n}Jͷ.0.'s>-|"-?5ȩL8ҦAp>1mDWKWW15iD0*M,Hrzw'4Eji2|fU(Ec e|yqN9d'H8@.e ItԺY-J\YF^mjȫ|H@+vx=7c"icrֈ,S^bV2C&xjF cRs'S:lKmzv3v]Cnd.`B䠷fRouaV(h0^y)hÌ$ - xh2xҲW T7w >Ak&!Jm1Y]JI3#QZ,Q:sRsixL iR-H6KYUb2rf|݃vwkL"{.}f|VqR}1-[7kf|T_ñVdtҌDI<[~3^Kf|H N_F;k _ٵgziTt vzi4٤Y%bAbYngKQ kjKs cW,o94}rl}SN1`1,>qS{Zh' mAHAS5\T?"qz(E&el *2O3F$&rInTX4 [[֚Jˮb@ЏoݫIKrYAN%a"53\>./Gskef/ w{/rĸ>?_./E4 /75j?| $.i.B(]S:(i %6vfW/A#?ːKٛ}žqE;lr(Ϻf@5mf`llYK#C5eΛ'ַCq;v(]CCzqčP iq kVO.dtuWC2Wvʲ\Rъ䑁3dv~GmMhۚ9-XAw_u۝qFg{9osPKxi c^/j2'$3!Am믚<мᡸmHC++1ވ3Aa`RFͩ6I5pmz-jh8K =mX:*hc;Nݗ&omU&]בס40>ϟPm9&9$D!7TfrD:Lm)ƅ`""O2].ד)M&+jUC(tDGyj%OkזLT(Mz(GyjўPz(f*?;&&W>.L |>q7tz PʹT6}s2ӻ9,c/+8} [Xol Yj8yn9/WP?-'>$' /u1yz'Wǚƺc].XVW*bSb q/Q}^ZvYW, j=>R-]xB14 j(CiA5RJBġTYQьBFR-g̀1&6fv"zw!1Y+憘Ejw0c>iTF/:bBUJmPވhވo߾mYLmԞWE]n.JeaQK< [V'rl}ԥ_\׫>i!>)$Y,..=S$Z-KRmb Bi3-bNI9VBmD /{c "S"AhhՉHʹ.Db*f?I7եXjv\Ƶrj]'zo_ʼ'3a Y_s<۸z6Cu9_lH?V~g OB1NU/7!fwAdޅ~ԫ?go o*& *g߬"|77%37_"VXqPbRhm4!I>C,4 ?EZv?}88*EK?L+ʷUz:F{mҷL$ta*eHa9NSS/sN4 3,Uӷ },b[ZS'51>[.LH-ݞuMGG~9f# yk[{7Pb0sb3h"x SFIљFb-wybE :Hgɒ'w(G eA # E f3awύ_aK*w.@oEbUrܷ\\J2Ez/yJ[eCL?F)l", 9IP862 #| U^/?|/ފ-6quIMů-_{>|Yf撉\}gSUӾ$]K|clwsMz@*2+(ɕe Vxt @6ζ|7j#Z?__Pmá 2aJ) yv/+AD)[o^88*ں4袥lu@n+I_?Xm݈B!Pr hYvR{ع, n?.ry~pLÅXiY?Ĝ(3K̉K1-k2zDMB)Q:uh%n_Ȫ_>+o6LmKt?@!M'd墮_bx3w3 z +J쯯թ VS4 `;vGo"1J(*ugF0]%#l7CEw&"8&H ) $V(䄅,͑!sy G^#@m~ar @/Ο}jSkwXbo_/#,YY̍]s ^߽LO "ګE= j2'kя`/ T7r7y!l<.{~;pS#/!_i肓,)UYحAeղg~^*nie9~5]X=|X-=Ll{h0kITkB[&`Ús‰uj؉hS/'1!Q[p])6uZF?)#r窱Z5;¦ք#G/#kI40.X=,HF? ֙$^RvY NpK++ROԗxHf#_1Lj2P4kާOLkcw5b[xBmmׯ6#LMg4dW%/R%BP*4Fx T;ޱ6TY 8iP [Zm&\\j{e=f$  4X?of%ey3oFPͬ#[Rri^ᑜd8Z;t &(Ӎcc =Z#E%cK;DЉYF(R9>"uD`*SIjRV<RSbSjڤag h em`fRc`\R^D|g6W"hi/^^EWCYĝ斢vc!^zc"BJ<.)od[p#6j![V([6bTJ#,zaM'θ 8,M1ENӆ`q,@A.r'>IvG?LI-%?ZD¶:Pk*L>T4CrC)jk5oFPcJSQ f-7q#Yyc4ŻuP0 bgB4MH!y!{<*I cfs1^6(rA`";y g}e_mhW[λW^\?#{ijNuHn^;Gv2nQep8Fb;ո[}sjUgLrX]]C txNzPI 4*]ҡEɑthyrCT.MTVZAbu'XLVeL8t[Й.h,&ўxdrJ$O*s"NiԒK=g;OVXr/=88i.ʯSb*Bh'Th} E'56JU-;کyGaj8Ղ/;Kv,5 RVdagAPzoA.4f:};vxl'T6YZHFe}!ŰKZB[ۅC0"8=fcW I1I<QxR&5ٳ" pNP"oƫ6Lw.uP(ywomhym4cM#-4NpC^ح蔣rru T*̩Ð|О0 ɛDh+сd* KJtwb+*NΊ4znV$f]GF FIP0[OD-TfJD4}Ӎ rZ]# ZuM4{K$G6g$>b$J먐gS-Rmg3@ wPBiJZ @ * [f[eYa$*fJ\H$JNhǍA<71J_,:ƌ-,cb@F}ޣ3㻽G+PN^}&09T}#p 1g1FB T&0*DGWh4S2T)|r݌"E)u}2n`S.#OK*8+G` O`E0HF)EbQJ1G"OdE NL)0kHwˠl5̬{SS8W]BoW]Zv߀WDω wsd,]XNų>2hWFCh_z'k|ܽt/,,Ӫ0Pa4@m^_{YTnL݇xvVɝ__>օ9E_ wخ|_L#Pq<8|-;ŏwٗ=]i?nFvBk>_yք^5b*Rp|d/gD"f83 r8(fh~G8:E7K;${zbW樖|E{zs{k-%)_܏Y?=S?fAgɧ9j:/ <_ogŵSn`xy7n+'N'Y2B<߃__&;?쩧RvI㼡쬧ONsJ۴npDU%S;X3Y `[bB ʧh9qnX7Eu*m1w h֭{#Ruc|StmNB 9!ʲA𕌰O\ko o};nWzFIp{ua)l+ mQj~r7>(Ӻ_ vOeiHUBsģQ@4YpNʀSNHAEr QIG% fj鯯V пGֵևL!r2[k{;O^.P0͝R 尹Waas+hzm">`IriwE+FgsNMꄈ6A2ӎK@,R.(Tui49Rd:ƝЉf%NϬQ,$(!(MFBn_=/M9Ù`~ wZ=Z|*AzS>9${%{$D0pYآy|tRP,lQ'J;)c^L;Cz{||ugr8Qֈ\I^pkdp;_ .l*J/wJbOɭLT2|)^o3)+GF{ KT/^QOH9s_B2`SÏm䆫s VV7pY'>T*)-h{kh=0Z(Ԇ!(=6/dO%rb5ȅw9k+Mg5#BHPB;M"害"bЮu} nZgQE\'45n2%pKAII' ^ / -7 T\-w>[yɵec|],zBZ?VYW)J+Y+9ܔJ詶:;{?ݬ%*RJY"\hIP5rс %VSľNZre1&oԉ?NMmpyrZ']̜+o~6DdPb~\ =q9#n(^w -O+ǫ+w]!tH*rG\~ ,ppR+ɹ]̰,>i6%ۚeX3F:Ma:2k6)כnҍhP;ZHpcKP!9YHÁ<J)aI"'@Vm>QZSJ(U6yd5ardy/l!ΜR7iǾ1g|>P>,$H,$V,FϵE g#)M iϢbG53l+ w3;k=[@G0h ǎ:&w_VTiW:?2# ca1l~^ te_eU]\;USK2޵>E/I.'- *lUn|qznm3s6߯AiG$ER3f(C~I6^Ǘ<ƥw\J7]^RzĎKݕvrǥӣKq7t\ʒh,؅f5*,K&7B9sӘٓku~\YӚ< 'Ǧr}SB,iq'N#O=u@lZF3bCry)$rH :}ߑݦLMKnC CDΞ#Z4q'n}ynS"ti$y) `}YJag ܎r-:}*ѣE4M߃V >^Aw&&Gxbk7ǤS# d a.SwX2}1Z,gϒ`dJ'\!k_|7nw}g !m&q*f%rp]NzY~nys2֨ԑ`L$i TU(=*%2S$f Aj`+K.N#V/ DkJP9eҊȽ~K1z+*QduKCmIےf2%3綤GjK* )Eˆs|"sl>hWX+Pg YLͲgO x+{)ifa-JOt̰OǡӇeq- 8v P*YJIyAJV8$XDxZ}AˠPG UT= ,00@+nc*Da+e*>OK. g mJovqLG* MOhۉcmM[ow&IGO䴭ӛ4ոIJz85mkM i[yEC6N91Q#g&]b@m _xԝ!'Gkơ\Kk#sB`Hp]/>N뙶 w1QiϗPbYX~8m+ei['=+B>T_h䩟cN1>F$1X!n,,2cȃ6DB@;1)++( ,/ JkqIvD"tDi3Nfʵ%tڪVymڂĻK`$o-